Michael Grieves - Digital Twins, Simulation, and The Metaverse - 2024
Michael Grieves - Digital Twins, Simulation, and The Metaverse - 2024
Abstract This chapter presents a brief introduction and history of simulation, Digital
Twins and their types and replica twins, and the origin of “metaverse”. Underlying all
these technologies is the premise that information generated by these technologies
is a replacement for the wasted physical resources in human goal-oriented tasks.
The chapter then provides the characteristics of a Digital Twin-oriented metaverse.
It applies the characteristics to the different DT types. It concludes by discussing the
evolution of Digital Twins in replication and prediction that will see Front Running
Simulation as our crystal ball into the future. AI is predicted to play a major role in
making this evolution possible as an assistance to humans but not a replacement.
1 Introduction
Digital Twins (DTs) are a twenty-first-century concept that has enjoyed an expo-
nential growth of interest over the last decade. DTs originated as the underlying
component of another twenty-first-century concept, Product Lifecyle Management
(PLM). PLM represented a change from a functional-centric approach where each
function, engineering, manufacturing, operations, and support had siloed its data
and information to a product-centric approach where every function populated and
consumed from a common source.
    That common source needed to be DTs. DTs took advantage of the exponen-
tial increases in information technology to implement this lifecycle-based, product-
centric representations of physical products and artifacts. This DT model was
M. Grieves
Digital Twin Institute, Cocoa Beach, FL, USA
e-mail: [email protected]
E. Y. Hua (B)
The MITRE Corporation, McLean, VA, USA
e-mail: [email protected]
originally proposed for automotive and aerospace. However, DT use has been
proposed for almost any physical product or artifact that exists in the physical
world, both tangible and intangible. The chapters in this book include DTs for
industrial and manufacturing [1–4], nuclear reactors [5], health care [6], airports
[7], semiconductors [8], power generation [9], and batteries [10].
    Humans have always had physical representations of different scales and fidelity
of products and artifacts in the form of Replica Twins as described later. DTs are a
digital representation of counterparts that exist or are intended to exist in the physical
world. Unlike atom-based objects where all or none of the objects exist, the bit-based
data and information of DTs allow for granularity of representation. The fidelity of
that granularity and the intervals needed to synchronize the DT and its physical
counterpart will depend on the use cases that create value. In order to understand
how value is created, it is important to understand what information does and the role
information plays with the different types of DTs throughout the product lifecycle.
    DTs have gotten a significant amount of attention from both academics and
industry. However, equally important is a digital representation of the physical envi-
ronment that the DTs are or will exist and operate in. While it is necessary and
useful to have the data and information about the DT and its physical counterpart,
it is equally important to understand the forces of the environment surrounding and
affecting the DT and its physical counterpart. This means the creation of digital
environments or digital spaces1 that multiple DTs from different sources can inter-
operate in. These are what we are calling Digital Twin Metaverses that this chapter
will describe. This represents the next wave of evolution for DTs.
    While defining and exploring DTs and DT Metaverses are important, this evolution
will feature the important capability of simulation. Simulation allows us to predict
and anticipate the future, at least probabilistically. We have had simulations for as
long as humans have been in existence and could think. What is novel is that we now
have the technology to simulate outside of human minds with DTs. This chapter will
set the stage for advancing this evolution.
2 Simulation
1We will use virtual and digital as synonymous here. There actually is no true virtual space. The
virtual representation is always instantiated in atom-based physical material. In humans, it’s in the
carbon-based matter of the brain. In digital computers, it’s in the silicon-based matter of digital
processor and memory components.
Defining, Exploring, and Simulating the Digital Twin Metaverses                         3
Since man began to think, he has performed simulations. Man has used simulations
from the beginning of human existence for a wide variety of tasks. Simulation has
been used for planning, assessments, training, scenario generation, risk assessment,
experimentation, and even entertainment.
    Take the example of prehistoric man hunting game. Prehistoric man ran through
different scenarios in his mind of what he, his fellow hunters, and the animal would
do. For example, the hunter mentally simulated what plan would predict running a
mammoth off a cliff. He ran through various simulations of which hunters needed to
be where, what actions they needed to take, and what reaction the animal would have.
He then selected the simulation that he believed would have the highest probability
of succeeding.
    He shared that simulation with his fellow hunters by tracing his plan in the dirt
with a stick. His fellow hunters watched this simulation unfold over time as the hunter
traced the stick in the dirt, showing the movement of the hunters and their intended
prey. It was crude. It was primitive. It often didn’t predict the intended outcomes.
However, it was simulation.
    The history of the military is intertwined with simulations. Soldiers throughout
history were trained in simulation exercises. Tzu [12] writes of convincing an emperor
of his ability to train troops by doing a simulation using the emperor’s harem. The
D-Day invasion of Normandy Beach was planned via simulation in a Scottish harbor
that had the characteristics of Normandy Beach [13].
    Over history, humans developed more stylized simulations. Early Greek plays
were simulations of what would happen over time when certain events took place.
Later, in the Middle Ages, written stories were simulations [14]. In the 1900s, movies
came about and provided much richer simulations that could be shared by many more
people.
    The arrival of computers in the last half of the twentieth century advanced the rigor
and robustness of simulations. These simulations were mathematically oriented and
could be quite complex in terms of calculations. However, these simulations with
applications like GPSS were mathematically abstract, and visualization was limited
to reams of numbers on paper output [15].
    Fast forward to the twenty-first century. We can do very rich detailed simulations
that provide photo realistic visualizations of the simulation of physical objects in their
environments. These simulations reasonably mirror the changes of their physical
counterparts when subject to the same forces.
As noted above, simulations are about predicting possible future outcomes. There
are two mechanisms that simulation uses: causation and correlation. From a system’s
4                                                               M. Grieves and E. Y. Hua
view of causation, we have inputs. We know those inputs cause specific things to
happen in our system. We then get outputs. It is deterministic.
    In the correlation model, we have inputs. We don’t know deterministically what
happens in the system, but we get outputs. Even though we don’t know what happens
within the system, we do know that there is a relationship between the inputs and the
outputs. Varying those inputs will result in a varying of the outputs that maintain a
correlating relationship. The correlations may be very strong, close to 100% or very
weak, close to 0%. We can use strong correlations and ignore weak correlations.
    We may have a correlation of 100% or close to that, but we may still not know why
or how the inputs result in the outputs. These correlation models can be useful even
if we have no idea why one input variable would correlate with an outcome variable
[16]. Note that we are very uncomfortable and should be questioning correlations
where we cannot theorize some relationship between correlated inputs and outputs.
    Simulations can use both methods, causation and correlation. Simulation driven by
causation will use formula or algorithms to take inputs and derive outputs. Simulation
driven by correlation will take data that they have from previous input/output data
and apply that to the input data that they are attempting to determine outputs for.
    Causation will give deterministic outputs, while correlations will give proba-
bilistic outcomes. However, it is important to note that for systems of high complexity,
there may be other unknown causal variables that will affect the outputs. We may
be unaware of these causal variables or that what we believe are causal variables are
simply correlated with this unknown causal variable.
    Many of the computer-based simulations in the past were driven by formulas or
algorithms. The complexity of these simulations was such that this was sufficient for
the outputs that were needed. The simplest version of this is y = f (t). This could be
an object moving in a straight line with a constant velocity. The variable “y” is the
location at any time “t” produced by the function. While this is not usually thought
of as a “simulation”, it meets the requirements of the time-evolved definition. These
functions or formulae can be very complicated but are causally deterministic.
    However, as we get into more and more complex systems with numerous variables,
some hidden, we will need to use correlations from big data that we are collecting
from products that are in operation. As we collect more and more data from products
in the field, Bayesian-based probability models assisted by Artificial Intelligence
(AI) become more and more useful [17].
    Using causation when we have it is highly preferred. However, it needs to be
remembered that causal models are conditional and, for complex systems, may not
reflect or predict accurately or capture all causal relationships [18]. Probabilistic
correlation models can be highly effective in predicting overall successful outcomes.
We can also use a hybrid approach which combines using both causal and big data
correlations [19].
    Going back to our prehistoric hunter, if he was successful 50% of the time in
running a mammoth off a cliff but not knowing exactly why it worked, he and his
tribe were well fed.
Defining, Exploring, and Simulating the Digital Twin Metaverses                                      5
FRS is a specific form of simulation. FRS was introduced by Grieves [20] and is
shown in Fig. 2. FRS is a simulation that predicts future states using assumptions
based on physics and/or data. However, instead of the initial conditions of the simu-
lation being arbitrarily set, the initial conditions are taken from the state of actual
conditions in the physical environment. At every new time zero (t0 ), FRS simulates
future states and attaches probabilities to those future states. The future states that
are the concern of FRS are states of adverse events. Adverse events are events that
waste resources hindering or preventing us from completing our task goals.
    There are two versions of FRS. The first is FRS using inputs from only the physical
product itself to predict future states. The second is that FRS uses inputs from the
physical product itself and the environment to predict future states. FRS acts as
crystal ball into the future.
    The specific conditions of FRS are:
. A simulation that contains behavioral assumptions of a corresponding physical
  product’s future states based on physics causality and/or data probabilities, usually
  Bayesian based.
. The initial conditions of each simulation at t0 are taken from the current state of
  an object in the physical world and, optionally, the environment that surrounds it.
Humans and non-human life, which for wont of a better term we will call “nature”,
have two different approaches to existence. Nature tries all possible combinations
and lets the environment select the winners. Nature can do this because its only goal is
survival of the fittest, and it has effectively an unlimited time horizon and resources.
Nature also does not care about individual living organisms.
   Humans, on the other hand, do care about individual living organisms, especially
their own. Humans do not have unlimited time horizons and resources. Humans also
have other goals besides survival. The human approach is to be task goal-oriented
and to accomplish that task using the minimum of physical resources, time, energy,
and materials.
   Given that Is the case, the human approach means that for a goal-oriented
task, humans employ sophisticated thinking capabilities utilizing data, information,
knowledge, and wisdom (DIKW) [21]. This relies on the fact that the expended
physical resources used to perform the goal-oriented task can be divided up into
two categories. These two categories are shown in the left bar in Fig. 1 [22].2 The
lower part of the bar is the minimum amount of resources that is ideally required to
2The figure in the book had a third category, Execution Inefficiencies, where we know what we
need to do to eliminate the waste but don’t have the technical capabilities yet to do so. That category
was dropped in later versions in the interest of simplification.
6                                                               M. Grieves and E. Y. Hua
complete the task. Everything above that is in the second category of wasted phys-
ical resources. We can apply a cost function to this in order to value the time, energy,
material of the physical resources needed to accomplish the task (Fig. 2).
   The right-handed bar shows the role of information in task accomplishment. The
amount of physical resources necessary to accomplish the task in the most optimal
fashion stays the same. Information cannot replace the minimum amount of physical
resources necessary to complete the task. However, information can replace the infor-
mation inefficiencies or wasted resources (C w (x)) over and above that. For purposes
of illustration, this figure shows information as replacing all the wasted resources,
but for human endeavors, this will usually not be possible.
    The issue with this costing function is that the cost of generating information
from data3 does not come in a unit of measure, like physical resources. We can use
time and physical resources needed to develop this information as a proxy for a unit
of measure. In the past, that proxy has always been human resources and physical
materials. Today, the proxy commonly consists of computer hardware, software, and
energy in addition to human resources. As represented in the formula, the assumption
under which this model holds true is that the cost of developing this information is
less than the cost of wasting resources over all the times the task is performed.
    Simulation is a method of developing information. Humans have no interest,
let alone the resources, in trying all possible combinations and letting the environment
dictate which one is successful. Humans want to simulate the possible ways of
obtaining their task-oriented goals and then perform the task using the method that
minimizes their scarce physical resources of time, energy, and material.
    Throughout all of human existence, until very recently, those simulations involved
only the computational capabilities of human minds. The development and rapid
advancement of digital computing bring a quantum leap in simulation value
for human goal-oriented tasks. The more accurately predictive and cheaper that
their simulations are, the greater value humans will obtain in completing their
goal-oriented tasks while minimizing the use of physical resources.
3 Digital Twins
3 There is much confusion about what is data and information. We have a functional perspective.
Data is a fact or facts about reality and the input to create information: We collect data and process
it to create information. Information is a replacement or substitute of wasted physical resources:
We use information.
8                                                                    M. Grieves and E. Y. Hua
There are physical objects that are “twins” in the sense that they are simply indepen-
dent duplicates. What we are interested in is 3D physical objects that are intended
to represent a specific physical object.4 Since the term “Physical Twin” (PT) refers
to the physical object in the dyad of Physical Twin–Digital Twin, we will call these
3D physical objects Replica Twins (RTs). We will use Replica Twins (RTs) in the
sense that there is a unique physical object and a replica physical object that can be
at different scales and fidelities.
    RTs, even if primitive, rudimentary, and abstractly shaped, have been used in
human endeavors for all of human existence. RTs have been used at all scales, from
small models to full-size replicas and all fidelities, from exact replicas to simple
representations. While barely three-dimensional, prehistoric man sketched in the
dirt with a stick a representation of a mammoth, the cliff that the mammoth needed
to be driven over, and the positions of his fellow hunters. While military sand tables
date from the 1800s, equivalents date back to ancient Greece military use and most
likely before.
    Architecture has used RTs from earliest times all over the world. RT artifacts
have been discovered dating from at least 6000 BC. RT model making was prevalent
in ancient Greece. The making of RT models to represent actual physical buildings
existed throughout the world in all cultures [24].
    The RTs were even dynamic and not simply static. Watch any movie about World
War II. It will generally feature a table with a geographical map that people will move
around representations of military and naval forces. As dispatches come in, people
move these representations into different geographical positions so that commanders
can assess and plan their next strategic and tactical moves.
    But it wasn’t simply scale models that were replica Physical Twins. Full-scale
RT mockups have been created and used. Full-scale RTs have been used in military
preparation as long as military engagements have existed. As noted above, the D-Day
preparation included exercises on Scotland beaches that were the replica physical
“twin” of Normandy Beach.
    As discussed in the simulation section above, we could argue that since very early
times plays, and then later, in the twentieth century, movies have used full-scale RTs in
a form of simulation. Plays and movies have created exact replicas of existing physical
environments and then “simulate” activity within those environments. For example,
an exact replica of the US White House Oval Office appears in an innumerable
number of movies and television shows.
    For equipment and vehicles, RTs were used primarily in development. However,
RTs were also used to resolve issues with operational equipment and vehicles.
Airplane manufacturers used replica twins to recreate and troubleshoot reported
4Obviously, there have always been 2D representations of physical objects, such as sketches,
drawings, blueprints. However, we wouldn’t call them “replicas”, as humans must do much mental
work to visualize them even poorly as three dimensional. They are more accurately described as
abstractions.
Defining, Exploring, and Simulating the Digital Twin Metaverses                                9
problems with their airplanes in the 1930s. When problems were reported with auto-
mobiles, it was standard operating procedure for the engineer working on the problem
to find an identically configured automobile to try and recreate the problem.
    As we developed electronics, we could make these RTs dynamic on their own.
The company one of the authors worked for in the 1970s, Lear Siegler Corporation,
had an F-16 flight simulator in its Grand Rapids Instrument Division. It was physical,
not digital. However, dynamic flight simulators date back to the 1930s with the Link
simulator [25]. There have been dynamic replica Physical Twins of nuclear reactor
control rooms for training and emergency exercises for over 50 years [26].
    The Apollo program is often cited as the first use of Digital Twins. That myth is
still being perpetuated today.5 The common reference is to the Apollo 13 mission,
where the myth is that its “digital twin” was used to bring the crew safely back home
after an almost catastrophic malfunction.6
    The reality is that the digital capabilities of the most powerful computer main-
frames of the Apollo days were extremely limited compared to today. Main memory
of the most powerful mainframes of the era was in the 16 MB range. The Command
and Lunar Landing Modules had a miserly 2K of main memory. The extensive trou-
bleshooting on earth was done with a series of RT capsules that had no “digital”
aspects.
    Replica twins have been in existence throughout humanity’s history and are still in
use today. Replica twins have been abstracted in such representations as dirt sketches
and sand tables. Replica twins have been realistic scale models such as buildings and
even cities. Full-scale replica twins have been used to prepare for and track military
engagements. Replica twins have existed dynamically as in the Link, F-16, nuclear
reactor control rooms, and the Apollo space capsule simulators. The advent and rapid
development of digital computers enabled the logical next step of moving “twins”
from physical replicas to digital ones.
5 This can be independently verified by doing a search of academic papers using the keywords
“digital twin” and “Apollo”.
6 Apollo 13 might be the most amazing malfunction recovery story ever. One of the authors had the
privilege of meeting the Apollo 13 Commander, James Lovell, and hearing first-hand the amazing
story of two astronauts sitting on what was basically a couch in Apollo 13, lining up the earth’s
meridian vertically and horizontally perfectly on a reticule etched on Apollo 13’s window so that
they didn’t burn up or bounce off into space at re-entry. However, the “twin” involved in working
the problem on earth was a replica twin capsule simulator, not a digital twin.
10                                                              M. Grieves and E. Y. Hua
its physical counterpart needed for use cases. At its ideal, any data that could be had
while in physical proximity of the physical counterpart could be obtained from its DT.
The characteristic of prediction is that the DTs would causally or probabilistically
predict the future states of their physical counterparts.
    The origin of the Digital Twin model is well documented in a multitude of
academic papers and industry articles [28]. Figure 3 is the first version of the
Digital Twin model that was presented at a Society of Manufacturing Engineering
(SME) conference in Troy, Michigan in October 2002. The presentation was entitled
Completing the Cycle: Using PLM Information in the Sales and Service Functions
[29]. It was about using Digital Twins in the operational and support phase of the
product lifecycle when there was both the physical product and its Digital Twin.
    The model in Fig. 4 was refined a little later that year to emphasize that prod-
ucts existed in real and virtual spaces. This version was for a meeting of industry
executives, automotive software providers, and academics from the University of
Michigan. The meeting was to explore setting up the Product Lifecycle Management
Development Consortium (PLM DC) at the University of Michigan. Because of the
automotive industry attendees, the focus was on different product lifecycle phases
than the SME conference, namely engineering and manufacturing.
    Both presentations were about the new discipline that was being defined, Product
Lifecycle Management (PLM). As a result, the model did not even have a name, as it
was simply entitled “Underlying Premise of PLM”. It did describe that the model’s
purpose was “Information Mirroring”. Even though it was rudimentary, the model
contained the major elements of the Digital Twin that exist today.
    On the left side were physical products in physical space. These are the Physical
Twins (PTs). On the right side were virtual products, which we now refer to as Digital
Twins, in virtual space that corresponded to the physical products. The third element
is that there were communications between the two spaces and products, with data
from physical space and products obtained from sensing and IoT devices populating
the virtual space and products, and data and information coming back from virtual
space and products to be used in the physical space.
    These models also contained the sub-spaces as part of the virtual space, VS1 ,
VS2 , VS3 … VSn. The idea of virtual spaces was fairly new at the time, so this was
to emphasize the fact that while there was only one physical space, there could be
an unlimited number of virtual spaces.
    This model highlighted that there were two main functions that it implemented:
replication and prediction. Replication is the characteristic that the DT would possess
all the data of its physical counterpart. The products in physical space were replicated
by the products in virtual space. At its ideal, any data that existed in the physical
product was replicated in its Digital Twin.
    The characteristic of prediction is that the DT would causally or probabilistically
predict the future states of the physical counterparts. The subspaces were virtual
12                                                              M. Grieves and E. Y. Hua
areas where predictive simulations could be performed. The multiple subspaces were
indications that there was no limit as to the number of simulations that could be done.
    The model was changed during work at NASA. The graphics are better, courtesy
of NASA, when the model was used for a Department of Defense (DoD) conference
[30]. In addition, the model was also simplified, as shown in Fig. 5. It was felt that
sub-virtual spaces unnecessarily complicated the model. However, as we will explain
later, the original model better represents a metaverse model.
    While the model did not have a name originally, it did acquire some names shortly
thereafter. It was called the Mirrored Space Model first, and then shortly thereafter
that was changed to Information Mirroring Model. The Information Mirroring Model
name remained albeit somewhat obscurely until around the 2010 timeframe. At that
time John Vickers of NASA who was working with Grieves suggested the name
Digital Twin. The Digital Twin name was a replacement for the relatively strange
name that NASA was using, Virtual Digital Fleet Leader.
    The Digital Twin name was mentioned in a footnote in Grieves’ book on PLM
[31], attributing the name to John Vickers. Grieves used the Digital Twin name in
one of the seminal and highly cited Digital Twin papers [32], which noted that the
Digital Twin name was going to be used for the model from then on. Later, in a short
but highly influential piece, Grieves wrote an article for the Economist Magazine in
the GE Lookahead section that was subsequently picked up by the World Economic
Forum [33]. The Digital Twin was explained to the general audience in that short
article.
    2015 marks the beginning of an exponential growth in reference to Digital Twins
in academic papers, industry white papers, and websites. The uses of DTs were
initially proposed for aerospace and automotive. That has exploded to encompass a
huge swath of industries and disciplines: power generation, heavy machinery, smart
Defining, Exploring, and Simulating the Digital Twin Metaverses                        13
building/cities, oil and gas, ports and airports, archeology, and healthcare, just to
name a few. Doing an internet search of “Digital Twins” in 2018 resulted in only one
million hits [34]. A search of 2022 results in over 17 million hits. The number of
academic papers on Digital Twins shows a similar exponential growth (2015—295
results, 2022—17,100 results).
   As of today, the model in Fig. 5 is the accepted model of the Digital Twin. While
definition may vary and vary widely, images usually show Digital Twin representa-
tions that are fairly consistent in the representation of physical space and products,
digital space and products, and a two-way connection between them.
   The commonly accepted Digital Twin Model that was introduced in 2002 and
simplified to the one as shown in Fig. 5 consists of three main components:
. The physical products (PT) in the real-world environment.
. The Digital Twins (DTs) in a digital environment.
. The two-way connection between the physical and virtual for data and informa-
  tion.
   The third component, the connection between the physical and digital has often
been referred to as the “digital thread”. The connection is a two-way communication
connection. Data from physical products and their environment is communicated to
the digital environment and populate their DT counterparts for collection, assessment,
and response (CAR) [20]. Data and/or information (if action to replace physical
resources is proposed) is now available to the physical environment.
   The digital environment of the Digital Twin, referred to as the Digital Twin Envi-
ronment (DTE), requires that it has rules that are identical as possible to our physical
environment. We need to be assured that the behavior of the Digital Twin in the DTE
mirrors the behavior of its physical counterpart for the use cases we require.
   Finally, it is important to remember that when we refer to Digital Twins in a general
way, we are implicitly including all three elements of the Digital Twin Model. We
are not simply referring to the digital object that represents a specific physical object.
While it would be more accurate to use the term, Digital Twin Model, we simply us
Digital Twin.
There are three types of Digital Twins: the Digital Twin Prototype (DTP), the Digital
Twin Instance (DTI), and the Digital Twin Aggregate (DTA).
   The DTP originates at the creation phase of the product lifecycle. The DTP of a
product begins when the decision is made to develop a new product, and work begins
doing just that. The DTP consists of the data and information of the product’s physical
characteristic, proposed performance behaviors, and the manufacturing processes to
build it. The DTP should also include the necessary processes and practices to ensure
the product is fully supportable and maintainable in the field and to troubleshoot and
repair the product effectively and efficiently to keep it operational. As much of this
work as possible should take place virtually.
14                                                              M. Grieves and E. Y. Hua
    The DTI originates when individual physical products are manufactured. DTIs
are the as-builts of the individual products and are connected to their corresponding
physical product for the remainder of the physical product’s life and even exist
beyond that. The DTI implements replication. Much of the DTI can simply be linked
to the DTP. For example, the DTI can link back to DTP 3D model and only needs to
contain the offset of exact measurements to the designed geometrical dimensioning
and tolerance (GD&T).
    Pre-production physical product versions that are called physical prototypes will
have a DTI since these are actual instantiations of the developmental period. These
DTIs should be put to the equivalent digital tests and evaluations as the physical
prototype itself. Comparing the digital results to the physical results will increase
the confidence on relying on digital testing when the product is moved to full rate
production. Where there are significant deviations, digital testing can be improved
to converge on producing equivalent results to physical testing, with the goal of
dramatically reducing and even eliminating physical tests, except for a final physical
validation.
    The DTP will contain the manufacturing process, Bill of Process (BoP), and the
parameters associated with the BoP. The DTI will contain any variations that occurred
in actual production. For example, the DTP process may require heat treating within
a temperature range. The DTI will capture the temperature that actually occurred.
The data and information that is needed for the DTI will be driven by the use cases
of the organization. The digital testing described above will make it possible to test
digitally each DTI of its physical counterpart to enable a high level of confidence in
the future performance of each individual product [35].
    Because the DTI remains connected to its physical counterpart for the rest of that
physical product’s life, it will also contain the data from its operation. The DTI will
contain sensor readings and fault indicators. Based on use cases, the DTI will contain
a history of performance of state changes and resulting outcomes.
    The DTA is the aggregate of all DTIs. The DTA contains all the data and infor-
mation about all the products that have been produced. The DTA may or may not
be a separate information structure. Some of the DTA data may be processed and
stored separately for analysis and prediction purposes. Other DTA data may simply
be mined on an ad hoc basis.
    The bigger the population of DTIs, the more data that will be available to improve
Bayesian-based predictions. The DTA, which consists of the DTIs of physical
systems, is subject to model bias for its predictions. However, there are mathematical
techniques available for bias identification in DTs [36]. The DTA will also be the
source for Artificial Intelligence and Machine Learning (AI/ML) to predict expected
performance.
    In 2019, Grieves introduced the Intelligent Digital Twin (IDT) to explain the
role that AI would have in both assisting Digital Twins in their performance and in
dealing with the increasing system complexity and emergent behavior of products
themselves [37]. The view here was that AI was not a replacement for humans but an
augmentation for humans. IDT specifies four attributes for Intelligent Digital Twins
as active, online, goal-seeking, and anticipatory.
Defining, Exploring, and Simulating the Digital Twin Metaverses                                 15
   The characteristic of anticipatory requires that the Intelligent Digital Twin can be
constantly running simulations to look ahead into the future for its PT. That obviously
means that FRS is a critical component of Intelligent Digital Twins.
As defined above, simulations are one process that imitates another process. In the
case of Digital Twins, we require that the “process” we imitate is our physical
universe. Our DT simulations need to have as perfect fidelity to the laws of our
universe as we need for our use cases. The characteristics of materials and forces of
our physical universe need to be imitated as closely as possible. Simulation is what
is needed to implement DT prediction.
    The one exception to adherence to the laws of our universe is the cadence of time.
We are unconstrained by time [37]. In our physical universe, time is completely out
of our control. We cannot go back in time. The only way to go forward in time is to
wait for the next tick of the clock. Even then, we go forward only according to the
set time. We cannot slow time down nor speed it up.
    In digital spaces, we are time unconstrained and can completely control time.7
We can run our simulations at any clock speed. We can computationally go years and
decades into the future. We can also slow down time. We can break down actions
that happen in split seconds in our physical universe into microseconds.
    In digital space, we can even go back in time which we cannot do in physical
space. In the physical world, we employ forensic methods to attempt to determine
what happened in the past that resulted in the current present. In simulations with
deterministic rules, we can usually reverse the arrow of time in the digital world. If
we have been using traceability to capture state changes, we can simply step back
through the time frames.
    However, another fundamental advantage is prediction, being able to advance the
clock to see what’s going to happen in the future. This is a crystal ball that sees into
the future. With Front Running Simulation (FRS) described previously, we have the
ability to do just that.
    The assumption is that “simulation time and wall clock time can be kept in sync
using conservative and optimistic synchronization protocols” [38]. At every new
time t0 , we will be able to take the data from the physical world, i.e., replication,
and predict at least probabilistically what will happen in the digital world. This is a
tremendous opportunity to prevent the waste of physical resources by anticipating
7 I discovered this firsthand in the early 1970s. Even though I was only a sophomore in college, I
was a systems’ programmer for a computer timesharing company. We charged by the CPU second.
There was a meeting to discuss how to increase revenue. After listening to the staff provide ideas,
I simply said, “I can increase what a CPU second is”. That worked in increasing revenue until
some customers started running benchmarks and complained that their programs were taking more
elapsed CPU seconds and therefore were more expensive to run!
16                                                              M. Grieves and E. Y. Hua
and correcting adverse events, especially ones involving the safety of individuals,
before adverse events can occur.
The purpose and value of DTs are that they provide information that can replace
wasted physical resources. One of the three elements of the Digital Twin Model in
Fig. 5 is the connection between Physical Space and its Physical Twins and Virtual
Space and its Digital Twins. Data is sent from the physical products and optionally
the physical environment to the digital DTIs. The information that is created and
housed in digital space is used in our physical space. This information is created
by processing the data coming from the physical space and, as noted above, by
performing simulations.
   This information can take a couple of forms. It can be the result of humans doing
queries of DTs and creating information by using the result of the queries to take
action that will replace wasted resources.
   For example, a certain model fuel pump of a helicopter is being recalled and
replaced. The traditional method would be to identify the location of every helicopter,
send a mechanic to inspect each one, and replace the fuel pumps in the helicopters
with the defective ones. With DTs, a query would be run on all the DTIs. Mechanics
would only be sent out to only those helicopters with defective fuel pumps. That
results in information replacing the wasted physical resources of mechanics’ time
and expenses for the helicopters with fuel pumps that need no replacing.
   Information can also be created by routines that run on a constant basis in digital
space looking for specific sensor data patterns that, using physics and DTA data,
simulates and predicts adverse events and alerts humans to them. Humans have the
responsibility of deciding the actions to take to avoid wasted physical resources by
having an adverse event occur. This is a human-in-the-loop version.
   The information can also be routines that run in digital space that specify actions
to take when they find certain conditions. These actions are coded as commands
and sent to the PTs directly, without human intervention. FRS is applicable in both
simulation-based situations. All are the result of data coming from the physical world
to digital world and data and information returning from digital world to the physical
world to replace wasted physical resources.
4 Metaverse
The original view of the metaverse is substantially different from what is required
for the purposes of Digital Twins. There is no reason that there cannot be all types of
metaverses with different rules governing them. The rules can be completely arbitrary
and have only a tenuous connection with our physical universe, as was the case for
the Stephenson metaverse.
    However, we would contend that we can legitimately co-opt the term because our
metaverse required for DTs has more fidelity to the meaning of the words, “meta” and
“universe”. What we require is a digital fidelity to our physical universe with meta
capabilities. The metaverse that we require needs to have complete conformance
to the selected rules of our unitary, physical universe, although it can have meta
capabilities as described below.
    In order to move work from the physical world into the digital world successfully,
the metaverse of Digital Twins needs to enforce all the laws of our physical universe
required by the supported use cases. However, we need a metaverse that is specifically
tailored to Digital Twins. From now on, we will refer to this type of metaverse as a
“DT Metaverse” (DTM).
    As humans perceive it, our universe is a three-dimensional space populated by
objects. The DTM will have that same characteristic, a three-dimensional space
populated by DTs. This will allow humans to use their ontological understanding of
physical space to understand and operate seamlessly in the DTM.
    Figure 6 is an evolutionary model of the DT that has been proposed previously
by Grieves [27]. The claim is that we are currently in Phase 2, the Conceptual Ad
Hoc phase. This is the current state of DT evolution.
    In this phase, “the Digital Twin is an entity that we conceptually create from
disparate and even fragmented data sources. We use different existing sources to pull
data from. We start building correlations and even causations of data source inputs
to results. We build different simulation views and determine how well they map
to reality. We start to put manual processes in place to pull the data from different
sources, even if on an ad hoc basis, to create a Digital Twin view [27].
    The next phase in the evolution of DTs is the creation of DT platforms. These
platforms are envisioned to support multi-users and multiple DTs. DT systems are
possible in Phase 2. However flexible Systems of Systems (SoS) where multiple DTs
of completely different functions enter and exit will require platforms that support
interoperability.
    First will come platforms that support replication, Phase 3. Those platforms will
be followed by platforms for prediction, Phase 4. While not referred to as “DT
Defining, Exploring, and Simulating the Digital Twin Metaverses                   19
Metaverses”, that is in essence what these platforms are by our definitions here.
These platforms now have a name, DT Metaverse. They are consistent with the
long-term vision of DTs.
   The DT Metaverse has these key characteristics:
. There are multiple DT Metaverses.
. DT interoperability is a core requirement.
. All laws of the physical universe are implemented and enforced in simulations
  for all inanimate objects.
. The DT Metaverse supports both replication and prediction.
. Multiple immersive participants as avatars is supported.
. Meta capabilities are allowed for human participants as avatars.
. Time can be synchronous or asynchronous with physical time depending on use
  case and DT type.
. Cybersecurity is embedded in all aspects of the DT Metaverses.
The limitation of our physical universe is that there is only one. If the Many Worlds
theory is correct and multiple physical universes exist [46], we will only ever have
access to the single physical universe this version of us inhabits. In the digital
universes, we can recreate our physical universe at some level. Since we have no
physical restrictions on the number of computer-based spaces, we have no restriction
or limitation of the number of digital metaverses that will be available to us.
20                                                              M. Grieves and E. Y. Hua
The concept of a DT Metaverse implies that there will be multiple DTs in it. Other-
wise, the DTM would be no different than what we have today, which is a Phase
2 Conceptual Ad Hoc programming space. The ability to easily insert a DT into a
digital environment to monitor its interaction and behavior with other DTs will greatly
enhance their value and usefulness. Interoperability also implies intra-operability as
this will also foster component modularity within DTs.
    This means that the DT Metaverses are platforms. Platforms are hardware and
software infrastructure that provides underlying tools, services, and governance to
accomplish participative specific tasks and interactions [47]. Participative implies
that there is intended to be multiple users and multiple DTs. So, this means that the
platforms need to enable the interoperability of DTs.
    There are numerous organizations working on mechanisms for interoperability.
A common perspective for interoperability is to produce standards and ontologies.
While this works in the physical world, it is much more difficult in the digital world
because of the much finer granularity of data and information. As a result, to address
this, we have a multitude of standards, which means that we really don’t have “a”
standard.
    These other mechanisms that platform may employ include defining ad hoc
programming conventions, harmonization of programming conventions among soft-
ware providers, and a platform’s own middleware. A promising solution may be to
deploy AI. AIs may be able to explore the solution spaces between different DTs
and provide mapping and translations. As Fig. 7 illustrates, AI may provide both
intra-operability for components of a Digital Twin and interoperability for different
DTs in DTMs.
    Depending on the DT type and use case, the DT Metaverse will be useful for
its ability to support immersion for multiple people assessing the serviceability of
a new product. It may also be useful for staging and operating multiple products in
simulation from different vendors in virtually commissioning a production line.
Defining, Exploring, and Simulating the Digital Twin Metaverses                     21
There are two main characteristics that DTs need. Those characteristics are replica-
tion and prediction. Even if a product is being developed in digital space and has not
yet taken physical form, the environment that the new product is going to exist in
needs to replicate our physical universe for the use cases needed.
    When we have DTIs of existing physical products, we need to replicate the data
from those PTs that we require for our use cases. The DT Metaverses supporting
this will be on a spectrum. On one end, there is simply the DTI in empty space
that can be simultaneously and instantaneously interrogated no matter where its PT
is in the physical world. On the other end of the spectrum, we can have DTIs and
the surrounding environment it currently resides in that an immersive avatar can
experience.
    We need prediction both when we are developing the product and then when we
have a product that is in operation. We need to predict what forces that the product
produces will have on its own structure, its operation, and its surrounding environ-
ment. We also need to predict the impact that outside forces we’ll have on our product.
Prediction is done via time-evolved simulations and can be done based on hybrid
physics/data probabilities techniques. Again, the DTM will be on a spectrum from
simply having the DTI representation of PT itself to having the DTI representation
of PT and its current surrounding environment. This gives us the ability to be time
and space unconstrained with a probabilistic window into the future.
22                                                            M. Grieves and E. Y. Hua
The primary requirement of the DT Metaverse is that inanimate objects will behave
as they would in our physical world. In DTM simulations, if, as described above,
a DT motorcycle traveling at Mach 1 slams into a stationary object like a building,
that will generate a tremendous amount of force as defined by F = M * V 2 . That
resultant force will deform/destroy the motorcycle, the building, and the surrounding
location.
   Since the physical world is a tremendously complex place, it may not be feasible
to enforce all the physics of our universe. DT Metaverses will need to establish
the parameters of their adherence to our universe’s physical laws so that users will
understand what they can and cannot rely on. DT Metaverses will need to define the
use cases that they support.
The DT concept has always been intended that there be a social aspect in terms of
having many people collaborate in a shared environment [32]. However, it was orig-
inally intended that people be outside observers. When there is immersive capability
with people having avatars, we will need to carefully define and design the meta
capabilities.
    What seems to differentiate metaverse from virtual environments is that meta-
verses enable immersion and host multiple participants. Immersion can be done first
person using avatars and Virtual Reality (VR) and Extended Reality (XR) capa-
bilities. The metaverse is a 3D geographic space that contains visual DTs that can
be examined and optionally operated by the avatars. The requirement for avatars is
important for knowing what participants are present in the metaverse and to know
who is operating the equipment, so multiple avatars don’t try and engage with DTs
simultaneously.
    The DTMs will allow for multiple participants who have independence within the
metaverse. Participants are not constrained to a certain view but can move around
the metaverse independently of other participants.
    This does not mean that it requires an immersive participant to use and obtain
value from the DT Metaverse. Authorized outside observers can have access to a
DT Metaverse and benefit from both replication and prediction data. Whether or
not observers can be active participants in a DT Metaverse will depend on the rules
employed by that DT Metaverse.
Defining, Exploring, and Simulating the Digital Twin Metaverses                      23
While it is foundational that our DT Metaverse follows the laws of our physical
universe, it does not preclude having meta capabilities for our participants of the DT
Metaverse. As described above, they can take advantage of the capabilities that simu-
lation and virtuality afford. Participants are unconstrained by location if so allowed,
meaning that they can “teleport” from one location to another. Solid structures can
become transparent so as to see inside. As long as the inanimate objects are subject
to the laws of our physical universe, human participants can have meta capabilities.
    If the use case is that participant avatars are merely observers who simply want
to experience the contents of the DT Metaverse as would be seen in the physical
world such as the layout of a factory and its equipment, we can allow them meta
capabilities such as walking through solid objects. However, if the use case involves
people interacting with the DT Metaverse physical objects, their avatars need to be
subject to the physical rules like every other inanimate object. For use cases such as
physical factory walkthroughs or equipment training, avatars should not have meta
capabilities. That would defeat the purpose of these types of use cases.
The spaces below the DT Metaverse, indicated by VS1 , VS2 , VS3 … VSn, are virtual/
digital spaces, “sandboxes” is the term usually used, where different simulations of
our DTI could be performed. The initial conditions from the DT Metaverse would
all be the same, but other conditions and assumptions would be varied to understand
what different outputs would result. The main space and the sub-virtual spaces were
what was referred to above as the Digital Twin Environment (DTE) above.
    In this section, we will put together the concepts we have described above.
As stated above, the DT Metaverse will not be a singular entity. There can and will
be different DT Metaverses, depending on where we are in the product lifecycle,
what DT types we are working with, and what use cases we need.
    In terms of capabilities, as described above, the DT Metaverse must twin the
physical universe consistent with our physical laws. The DT Metaverse needs to
reflect what the physical world does. It needs to have persistent physical atom-based
material characteristics. This is called coherence that the physical world enforces
through atoms but needs to be specifically programmed into digital worlds [22]. For
example, if we lengthen a steel beam, it has to not only lengthen the beam, but it
must add physical weight.
    It must have persistent forces active upon atom-based materials. If a force acts
on a particular material, it must do it consistently across all the materials in the DT
Metaverse. We cannot have forces acting one way on part of the DT and in a different
way on a different part of the DT.
    We must have persistent time in each DT Metaverse. All the objects of our DT
Metaverse must have the same time basis. We cannot have some DTs speed up in
time and some DTs lag in time. There must be a persistent time basis in the DT
Metaverse. This requires Digital Twin interoperability. As we move different DTs
Defining, Exploring, and Simulating the Digital Twin Metaverses                      25
into a DT Metaverse, they all have to participate as if they were physical objects in
the physical universe. Interoperability of different DTs is a major issue that we need
to address.
The DT Metaverse of the DTP will need to support both replication and prediction.
In this case, since we are developing a new product, the replication will be of the
environment that our product is going to operate in. It will be a synthetic replication
that we will put our new product through its paces in. The primary focus of the DT
Metaverse in this lifecycle phase will be prediction of behavior over time.
    We will want to predict the suitability for purpose and behavior performance of
our new product under a variety of conditions. We want to predict our product’s
performance throughout its entire lifecycle. As has been specified, we will want
to create the product virtually, test the product virtually, manufacture the product
virtually, and support the product virtually. Only when we get the product as perfected
as we possibly can do, we want to move around expensive atoms to build a physical
version.
    The DT Metaverse will need to adhere to all the laws and rules implemented in
our physical universe for the use cases it supports. We want to test our products under
a range of conditions to see that they can measure up. We will want to put a new
airplane through years of missions that it will encounter. We want our pacemaker to
encounter a wide range of patient conditions and react with the required response
accordingly.
    We want to manufacture a product on a virtual production line. What we want to
do here is reflect the movement of materials, machines, and people. We will explore
different Bill of Processes (BoPs) to determine the “leanest” version. We will look
at the ergonomics of the factory workers to ensure that there are no health and safety
issues. As a byproduct, we can do our virtual commissioning of new production
facilities during this phase.
    In the support/operation phase, we will want to insert problems into our DTP and
see how we troubleshoot that particular issue. We want to disassemble and reassemble
to ensure that it will be easy and convenient in the field. We want to see how products
interface with support equipment such as refueling systems, to make sure that it has
been designed appropriately.
    We want our DT Metaverse to support interoperability. This will allow us to take
components from different manufacturers and make sure that they interface with each
other. Interoperability also means that the physical laws are enforced on a coherent
basis. When we fill the fuel tank in the wings of an airplane, the plane fuel gauge
will register the increase in fuel, the wings will deform appropriately for the amount
of new mass they have in it, and the weight of the entire airplane will increase.
    Humans will be able to participate via avatars. They will be able to walk around the
product, inspect areas of it, open doors and latches, and even operate the product. For
an airplane, they can get into the cockpit and fly the airplane through a mission. Using
26                                                                M. Grieves and E. Y. Hua
VR and haptic equipment, they can bring in a refueling hose, connect it to the refueling
port, and feed fuel into the airplane. They can participate on the manufacturing line
and move parts from one place to another. They can operate and change settings on
the digital machines that are producing the parts.
   These avatars can have meta capabilities, so they could jump from one place to
another or see through outside structures to see what’s underneath the surface cover-
ings. Avatar participants will have an indication as to whether time is synchronous
or asynchronous, so they can understand whether the simulation is being sped up or
slowed down.
5.1.2 Digital Twin Instance (DTI) and Digital Twin Aggregate (DTA)
We will consider the DTI and DTA at the same time, since the DTA is made up
of DTIs. When we move to looking at the DTI and DTA type, the DT Metaverse
requirement changes. The DT Metaverse is now driven by a Physical Twin. The PT
is at the mercy of our physical universe clock. PTs can only move forward in time
as fast as the clock ticks.
    Even though there is a unique DTI for every PT object, there is no reason that the
DTI cannot participate in multiple DT Metaverses. There may be different use cases
that are involved that the different DT Metaverses are concerned with. What will be
true for replication use cases is that in each of the DTMs, the DTI in that DTM will
be identical at times that are synchronous to our physical universe to the same DTI
in other DT Metaverses.
    The DT Metaverses for the DTI and DTA will support both replication and predic-
tion. At this phase in the lifecycle, there is actually a PT to replicate and the physical
environment that the PT lives in. Depending on the use cases, we need to populate
the DT Metaverse with both the PT data and the environmental data.
    Because we have sub-virtual spaces, we can engage in prediction. This is where
FRS comes in the play. At every time zero, we can run a simulation in the sub-virtual
spaces to predict the future of individual products or even a group of products or prod-
ucts as a whole. We can use the Digital Twin Aggregates to provide Bayesian-based
probabilities. For purely deterministic situations, we can simply predict outcomes,
such as predicting that unless a ship reduces its speed immediately that it will be
unable to stop in time and will crash into its berth. In other cases, we may provide
probabilities of different outcomes.
    As noted above, for real-time replicative use cases, DT Metaverses will require
that time will need to be synchronous and persistent. However, that will not preclude
use cases where it may be advantageous to drop out of real time in replication and
slow the happening of an event. This will give humans the ability to assess all that is
happening at a slower rate to make better decisions. Better decisions mean reducing
wasted physical resources to accomplish the goal by selecting the best future actions
to take.
Defining, Exploring, and Simulating the Digital Twin Metaverses                       27
    Obviously, our DT Metaverse will need to follow the laws and rules of our physical
world. While prediction of the PT itself is important, the impact that this PT will
have on the external world is also important.
    DT interoperability will also be important in this DT Metaverse. In a city-oriented
DT Metaverse, DTIs of various vehicles could participate. We would want to have
them interoperate with DTIs of roads, traffic signals, and other aspects of the envi-
ronment. This is so that we could have accurate predictions of what would happen
in this particular DT Metaverse as different vehicle DTIs moved about.
    We could use multiple participant immersive DT Metaverses. We might have
a factory machine that is having problems. Experts from all over the world could
congregate in the DT Metaverse of the factory to troubleshoot the problem and
recommend a physical fix. These experts could adjust parameters within the DT
Metaverse of the DTI machine that would then send those commands to the actual
machine itself. This is in essence merging the physical and digital worlds.
    Human participants could have meta capabilities. So not only could they be
looking at the DTI of the machine, but they could also look within the machine
to see certain elements of it so that they had a better understanding of what was
transpiring.
    Obviously, the machines would need to be instrumented to provide this data, but
that is fully within the realm of possibility. While dealing with a PT, there is no
other choice but to be synchronous. However, these avatars could step into one of
the sub-virtual spaces to speed up the process to see what was going to happen and
then transition back into the main DT Metaverse space in order to take action. This
looks very much like the original Digital Twin model in Figs. 3 and 4.
We need to consider the role that Artificial Intelligence (AI) will have in the evolution
to a DT Metaverse. It has been asserted that we are moving into an era of Intelligent
Digital Twins (IDT) that have the characteristics of being active, online, goal-seeking,
and anticipatory [37]. Grieves claim for DTs and IDTs is that AI will be an assistance
for humans and not a replacement.
   It should be obvious that as we move into DTM platforms with major requirements
for replicative and predictive analysis and simulation and ensuing adverse prevention
that humans will need help. That help will take the form of AI.
   In an above section, we discussed the difference between humans and nature.
While humans are task goal oriented while attempting to minimize physical
resources, nature tries all possible combinations via genetic mutations, letting the
environment select for fitness.
28                                                               M. Grieves and E. Y. Hua
    While the habitat of nature is the physical world, the habitat of AI is the digital
world or the DT Metaverses. Subject to computational limits and time constraints, AI
has the ability to explore the solution space, test it against the criteria for obtaining
goal success. AI can then select those solutions that meet or exceed the success
criteria.
    In order to evolve to Phase 4 with Front Running Simulation, we are going to
need AI capability. FRS will require massive amounts of data coming from DTIs on
a constant basis. FRS will then require selecting the assumptions and parameters and
running simulations of future outcomes. The probabilities of those outcomes will be
calculated from the DTA and its causal and Bayesian probabilities.
    However, it’s simply not enough to have probabilities of adverse events. We will
want to take action to prevent them from occurring. From the beginning days of DTs,
a key characteristic was cued availability [22]. Cued availability is a digital assistant
that assesses what is happening in context and offers information that replaces wasted
physical resources. Clearly an adverse event is an event that wastes resources.
    Cued availability is a role for AI. Some adverse events will have obvious solutions.
It is obvious that the remediation for the prediction that there is a high probability of
a component failure in few weeks is to replace the component before the predicted
failure. Other predictions of adverse events will not have obvious remediations. An
example is a robot glitching momentarily that will result in an assembly bottleneck
later that day. AI has the ability to propose alternative preventative solutions for the
different adverse events it predicts.
    This is clearly beyond human capability, computability, and mental bandwidth.
This will require the assistance of an AI agent with an enormous amount of computing
capability at its disposal. We may delegate some autonomous operations to AI
systems in FRS. However, we should always understand that AI should be an assistant
to humans and not a replacement.
6 Conclusion
resources in our goal-oriented tasks. We can use Digital Twins to replicate their phys-
ical counterparts so that we have simultaneous and instantaneous access to their data,
develop that data into information, and then predict their behavior via simulation.
   The idea of a DT Metaverse will allow us to continue to evolve this capability
as we move to Phase 3 and 4 DT platforms. However, the DT Metaverse will need
the specific characteristics that we’ve outlined, such as fidelity to our universal laws.
Having our products employ Front Running Simulations will be our crystal ball into
the future, with the opportunity of using AI to prevent adverse events.
   This may seem fantastical until we realize that it was only fifty years ago that
computer chip capacity was measured in thousands of bytes and that we recently
passed 100 billion bytes on a chip, with more exponential increases to come in the
next decade or so. With that in mind, the advances proposed in this chapter seem
very reachable and promise to create substantial value.
References
 1. Lu H (2024) A framework for the credibility evaluation of equipment digital twins. In: Grieves
    M, Hua E (eds) Digital twins, simulation, and metaverse: driving efficiency and effectiveness in
    the physical world through simulation in the virtual worlds, Forthcoming. Springer, Heidelberg
 2. Vickers J (2024) NASA’s Michoud assembly facility: developing a factory digital twin. In:
    Grieves M, Hua E (eds) Digital twins, simulation, and metaverse: driving efficiency and effec-
    tiveness in the physical world through simulation in the virtual worlds, Forthcoming. Springer,
    Heidelberg
 3. Malik AA (2024) Digital twins for robot systems in manufacturing. In: Grieves M, Hua E (eds)
    Digital twins, simulation, and metaverse: driving efficiency and effectiveness in the physical
    world through simulation in the virtual worlds, Forthcoming. Springer, Heidelberg
 4. Boschert S (2024) On the importance of simulation and digital twin for industrial applications.
    In: Grieves M, Hua E (eds) Digital twins, simulation, and metaverse: driving efficiency and
    effectiveness in the physical world through simulation in the virtual worlds, Forthcoming.
    Springer, Heidelberg
 5. Darrington J (2024) DThe first real-time digital twin of a nuclear reactor. In: Grieves M, Hua
    E (eds) Digital twins, simulation, and metaverse: driving efficiency and effectiveness in the
    physical world through simulation in the virtual worlds, Forthcoming. Springer, Heidelberg
 6. Dong Y (2024) Evolution of simulation and digital twin in healthcare: from discovery to design
    and integration. In: Grieves M, Hua E (eds) Digital twins, simulation, and metaverse: driving
    efficiency and effectiveness in the physical world through simulation in the virtual worlds,
    Forthcoming. Springer, Heidelberg
 7. Alexandridis K (2024) Distributed AI modeling and simulation for smart airport digital twin
    applications. In: Grieves M, Hua E (eds) Digital twins, simulation, and metaverse: driving
    efficiency and effectiveness in the physical world through simulation in the virtual worlds,
    Forthcoming. Springer, Heidelberg
 8. Moradian A (2024) Digital twins for sustainable semiconductor manufacturing. In: Grieves M,
    Hua E (eds) Digital twins, simulation, and metaverse: driving efficiency and effectiveness in
    the physical world through simulation in the virtual worlds. Editors, Forthcoming. Springer,
    Heidelberg.
 9. Mondoloni S (2024) The power and application of digital twins: a multidisciplinary perspective.
    In: Grieves M, Hua E (eds) Digital twins, simulation, and metaverse: driving efficiency and
    effectiveness in the physical world through simulation in the virtual worlds, Forthcoming.
    Springer, Heidelberg
30                                                                       M. Grieves and E. Y. Hua
10. Singh S (2024) Defining, exploring, and simulating the digital twin metaverses. In: Grieves M,
    Hua E (eds) Digital twins, simulation, and metaverse: driving efficiency and effectiveness in
    the physical world through simulation in the virtual worlds, Forthcoming. Springer, Heidelberg
11. Hartmann S (1996) The world as a process: simulations in the natural and social sciences.
    Modelling and simulation in the social sciences from the philosophy of science point of view.
    Springer, Heidelberg, pp 77–100
12. Tzu S (2000) Art of War
13. Allison D (2019) Scotland’s role in the D-Day landings [cited 2023 August 23]. Available
    from: https://www.bbc.com/news/uk-scotland-48266481
14. Barzun J (2000) From dawn to decadence: 500 years of cultural triumph and defeat, 1500 to
    the present. HarperCollins
15. Schriber TJ (1974) Simulation using GPSS. Wiley, New York, xv, 533 p
16. Cass D, Shell K (1983) Do sunspots matter? J Polit Econ 91(2):193–227
17. Korb KB, Nicholson AE (2010) Bayesian artificial intelligence. CRC Press
18. Pearl J (2003) Statistics and causal inference: a review. TEST 12:281–345
19. Guo Y et al (2023) Digital twins for electro-physical, chemical, and photonic processes. CIRP
    annals manufacturing technology, Forthcoming
20. Grieves M (2019) Virtually intelligent product systems: digital and physical twins. In: Flumer-
    felt S et al (eds) Complex systems engineering: theory and practice. American Institute of
    Aeronautics and Astronautics, pp 175–200
21. Grieves M (2024) DIKW as a general and digital twin action framework: data, information,
    knowledge, and wisdom. Knowledge, Forthcoming
22. Grieves M (2006) Product lifecycle management: driving the next generation of lean thinking.
    McGraw-Hill, New York
23. Popper K (2014) Conjectures and refutations: the growth of scientific knowledge. Routledge
24. Pillsbury J (2015) Modeling the world: ancient architectural models now on view [cited 2022
    September 7, 2022]. Available from: https://www.metmuseum.org/blogs/now-at-the-met/2015/
    modeling-the-world-ancient-architectural-models
25. New Link Trainer (1936) Aviation. McGraw-Hill Publishing Co, p 3
26. Furet J (1985) New concepts in control-room design: Worldwide, more attention is being given
    to improving the man-machine interface. In: Nuclear power and electronics
27. Grieves M (2023) Digital twins: past, present, and future. In: Crespi N, Drobot AT, Minerva R
    (eds) The digital twin. Springer International Publishing, Cham, pp 97–121
28. Holland S (2022)Michael Grieves, Father of Digital Twins, to headline at IOTSWC. In: iot
    insider
29. Grieves M (2002) Completing the cycle: using PLM information in the sales and service
    functions [Slides]. In: SME management forum, Troy, MI
30. Caruso P, Dumbacher D, Grieves M (2010) Product lifecycle management and the quest for
    sustainable space explorations. In: AIAA SPACE 2010 conference & exposition, Anaheim,
    CA
31. Grieves M (2011) Virtually perfect: driving innovative and lean products through product
    lifecycle management. Space Coast Press, Cocoa Beach, FL
32. Grieves M (2014) Digital twin: manufacturing excellence through virtual factory replica-
    tion (White Paper). Available from: https://www.3ds.com/fileadmin/PRODUCTS-SERVICES/
    DELMIA/PDF/Whitepaper/DELMIA-APRISO-Digital-Twin-Whitepaper.pdf
33. Grieves M (2015) Can the digital twin transform manufacturing. Available from: https://www.
    weforum.org/agenda/2015/10/can-the-digital-twin-transform-manufacturing/
34. Brown C, Singer T (2019) Defining and utilizing the digital twin. In: Digital enterprise society
35. Grieves M (2023) Digital twin certified: employing virtual testing of digital twins in
    manufacturing to ensure quality products. Machines 11(808)
36. Andres Arcones D (2023) Evaluation of model bias identification approaches based on Bayesian
    inference and applications to digital twins
37. Grieves M (2022) Intelligent digital twins and the development and management of complex
    systems. Digital Twin 2(8)
Defining, Exploring, and Simulating the Digital Twin Metaverses                               31
38. Oren T et al (2023) Simulation as experimentation. In: Oren T, Zeigler BP, Tolk A (eds)
    Body of knowledge for modeling and simulation: a handbook by the society for modeling and
    simulation. Springer Nature, pp 77–119
39. Grieves M (2006) Product lifecycle management : driving the next generation of lean thinking.
    McGraw-Hill, New York, xiii, 319 p
40. Stephenson N (1991) Snow crash: a novel. Bantam Books
41. Bell MW (2008) Toward a definition of. J Virtual Worlds Res 1(1)
42. Joshua J (2017) Information bodies: computational anxiety in neal stephenson’s snow crash.
    Interdisc Literary Stud 19(1):17–47
43. Rymaszewski M (2007) Second life: the official guide. John Wiley & Sons
44. Bohné T, Li K, Triantafyllidis K (2023) Exploring the industrial metaverse: a roadmap to the
    future. World Economic Forum and Univesity of Cambridge
45. MIT Technology Review Insights (2023) The emergent industrial metaverse
46. Carroll S (2020) Something deeply hidden: Quantum worlds and the emergence of spacetime.
    Penguin
47. Parker G, Van Alstyne M, Choudary SP (2016) Platform revolution: how networked markets
    are transforming the economy and how to make them work for you, 1st edn. W.W. NORTON &
    Company, New York, xiii, 336 pages
48. Richter S, Richter A (2023) What is novel about the Metaverse? Int J Inf Manage 73:102684
Contrasting Capabilities: A Comparative
Analysis of Simulation and Digital Twin
Technologies
1 Introduction
A. Wooley (B)
The University of Alabama in Huntsville, Huntsville, USA
e-mail: [email protected]
D. F. Silva
Auburn University, Auburn, USA
e-mail: [email protected]
comparison between DTs and simulations, this chapter emphasizes the synergistic
relationship between them while emphasizing their distinct characteristics and capa-
bilities, enabling informed decision-making and effective implementation of these
technologies in various industries.
    The chapter begins by defining the concept of a DT, introducing the 4R framework
[1, 2] to characterize and classify DTs based on levels of capability. Next, it presents
the definition of simulation and proposes an analogous framework to characterize
simulation based on its levels of capability. This chapter highlights that a simulation
can be based on reality, or it can be completely artificial, with the primary objective
centered on examining how an object or system evolves and produces outcomes over
time, using a predefined set of inputs. On the other hand, a DT fully replicates, in a
virtual environment, a physical element that exists or will exist in the physical world,
with bidirectional data flow connecting the physical and the virtual environments.
The main result presented here is a discussion of the similarities and differences
between DT and simulation, based on the capability-based classification frameworks
discussed earlier.
    In summary, the main conclusion is that simulation and DT share a common foun-
dation of modeling and analysis but diverge in their objectives and advanced features.
DTs, unlike simulations, offer real-time adjustments and autonomy, enhancing
decision-making and optimizing physical systems.
The concept of DT has gained significant attention from both academia and industry
in recent years [3, 4]. This surge in interest in DT is primarily driven by Industry 4.0
and advancements in technologies, such as the Internet of Things (IoT), Artificial
Intelligence (AI), wireless sensor networks, Machine Learning (ML), and big data
[5]. These technologies offer opportunities to seamlessly merge physical and virtual
environments. The growing interest in DT has also led to an increase of publications
about it [6, 7]. However, even academic publications about DT do not always agree
on the what the definition, characteristics, and capabilities of DT are, or should be,
resulting in a dilution of the DT concept [8–10], and causing confusion regarding its
practical applications and advantages [11].
    The concept of DT was originally introduced by Michael Grieves in collaboration
with John Vickers from the National Aeronautics and Space Administration (NASA)
[12]. Grieves’s definition of a DT consists of three key elements: a physical space,
a virtual space, and the data flow that connects these spaces bidirectionally. Since
then, researchers have been trying to establish a unified and universally accepted
definition that effectively captures the essence of a DT [6, 10]. Some researchers,
as exemplified by [13–16], focus on DTs applied to individual products, where a
DT represents a single machine or process. Conversely, others, including [17–21],
consider DTs in the context of processes, encompassing either a segment or the
entirety of the production environment.
Contrasting Capabilities: A Comparative Analysis of Simulation …                     35
virtual systems are connected and the data between them is bi-directional. Addition-
ally, the DT has a level of autonomy that is not present in the preceding capability
levels, enabling it to autonomously adapt, identify optimal strategies, and fine-tune
itself.
    A concise overview of the DT capabilities is depicted in Fig. 1.
3 Simulation Capabilities
analysis. The three most prevalent simulation paradigms are Monte Carlo, Agent-
based, and Discrete Event Simulation (DES) [39].
    In simple terms, a Monte Carlo simulation is employed to assess uncertainties and
potential risks by means of probability distributions [40]. Monte Carlo simulation is
often used to generate random variates that will be utilized in more complex analysis,
such as Monte Carlo Markov Chain models.
    In contrast, an Agent-based simulation models individual entities (agents) and
their interactions with each other and the environment, facilitating the modeling of
intricate systems with numerous interacting components [41]. In Agent-based simu-
lations, agents act in a decentralized manner, following a set of prescribed (possibly
randomized) rules that govern their interactions with other entities.
    DES, on the other hand, permits the modeler to generate or observe specific
events over time and it proves versatile for studying various system types and an
extensive range of outcomes [36]. DES stands as the most used approach within the
manufacturing domain [42].
    Extensive literature exists defining simulation and offering detailed instructions
for constructing simulation models. Many authors agree [35, 43–46] that simula-
tion models provide the capability of conducting experiments and analyzing results.
However, none of them provide a structured framework for characterizing and clas-
sifying simulations that would allow us to distinguish between the different levels
of capability or that can be directly compared to other technologies (by comparing
capabilities), such as DT.
    While numerous frameworks and guidelines exist in the literature for constructing
simulations, to the best of our knowledge, there is no framework that precisely delin-
eates the different levels of capability within a simulation model in a manner that is
comparable to a framework for DT, such as the 4R framework. To be clear, the capa-
bilities of simulations are thoroughly documented in existing literature; therefore, the
objective of developing such a framework is to consolidate these well-established
definitions and integrate them into a comprehensive structure that can serve as a
roadmap for understanding simulation models based on their capabilities, which is
analogous to the existing 4R framework for DTs.
    An analogous framework for evaluating simulation capabilities should provide
a similar assessment of the level of simulation maturity in an organization. This
would enable organizations to gauge their current simulation capabilities, pinpoint
areas requiring improvements, and develop a plan for future simulation investments.
Furthermore, an analogous framework would facilitate comparisons between simu-
lation and DT capabilities, allowing organizations to better align their simulation and
DT strategies.
    In this context, we present a framework for simulation, which we called the 4S
framework, that was initially introduced in our prior work [47]. It is designed to be
analogous to the 4R framework for DTs, allowing for the classification of applications
of both DT and simulation and facilitating comparisons between them. It is important
to emphasize that there isn’t a one-to-one correspondence between the levels in the
4S and 4R frameworks.
38                                                                    A. Wooley and D. F. Silva
Fig. 2 Systematic search for textbooks and articles. Figure adapted from [47]
each characteristic (C1–C6), it was found that certain categories exhibited significant
conceptual overlap. As a result, C1 and C2 were consolidated into a single category,
and C3 and C4 were likewise combined.
   The categories “Understand how the system works” (C1) and “Represent the
physical system in a digital model” (C2) were combined into a single category
because many authors combine understanding and representation of a system into one
step [35, 36, 43, 49–51, 53, 54]. C1 emphasizes the need for understanding a system’s
behavior to create an accurate simulation model. This encompasses the identification
of various system components, their interactions, and their interrelationships [58].
On the other hand, C2 involves the development of a digital representation of a
physical system. This process involves constructing a simulation model that faithfully
mirrors the characteristics and behavior of the physical system. To accomplish this,
the simulation modeler must have an in-depth understanding of the behavior of the
physical system, its components, and their interactions [33].
    The categories “Examine and analyze the system and its components” (C3) and
“Draw inferences about the system” (C4) were combined into a single category
because both involve the process of analyzing a system to draw conclusions about
its behavior. C3 entails the examination and analysis of various system components
within the simulation model to understand how they interact with one another. This
analysis encompasses the identification of relationships among different components
and their impact on the overall system behavior. Similarly, C4 is centered around
drawing conclusions about the system’s behavior based on the outcomes generated
by the simulation model. Upon the completion of the simulation, the modeler can
analyze the results and draw conclusions regarding the system’s behavior [59].
    After combining similar characteristics, the resulting framework comprises four
categories. To make it more comparable it with the 4R framework of DT, these
categories are defined as the 4S framework of simulation: S1 (combining C1 and
C2), S2 (combining C3 and C4), S3 (C5), and S4 (C6). To appropriately label the
four categories in the 4S framework of simulation, the most common and relevant
terms associated with each category were visually represented in a word cloud. This
approach helped to simplify and condense the keywords, making it easier to identify
the most important and frequently used terms. The word clouds for each category
are shown in Fig. 3.
    Following a thorough analysis of the word clouds, names were assigned to each
of the four categories in the 4S framework of simulation. To encapsulate the essence
of each category as a capability, it was decided to use action verbs when selecting
category labels. In the word cloud of the first category (S1), the most frequent terms
Fig. 3 Word Clouds for labeling the 4S levels. Figure adapted from [47]
42                                                                  A. Wooley and D. F. Silva
Fig. 4 Summary of the simulation capabilities (4S Framework). Figure adapted from [47]
Contrasting Capabilities: A Comparative Analysis of Simulation …                       43
requirements. In the computer modeling stage, the system components are repre-
sented in a virtual format and the simulation model is created. Proper data collection,
parameter selection, and understanding are crucial in this stage.
   Analyzing (S2): In this level, the data gathered during modeling is incorporated
into the simulation model. Input analysis ensures that the data used approximates real-
world behavior. Verification and validation are performed to ensure model accuracy.
In this phase, analysis is conducted to draw inferences about the system’s behavior.
   Predicting (S3): This level is where the simulation has the capability to predict
system behavior under various conditions. It allows for the investigation of alter-
native scenarios, performance measurement, and comparison. Developing accurate
scenarios is essential.
   Prescribing (S4): This level involves using the simulation model for decision
support and optimization. Decision-makers can identify optimal courses of action,
and simulation optimization or machine learning methods can be applied to enhance
decision-making and system performance.
       60
                                                   2
       50                                          9
       40          2
                                                   13
       30         19
       20
                                                   30
                  13                                              9
       10
                   6               9                              8
        0
              R0 (33%)        R1 (8%)        R2 (45%)         R3 (14%)          R4 (0%)
S1 S2 S3 S4
Fig. 5 Analysis of the systematic literature review. Figure adapted from [63]
   With the introduction of the 4S framework for simulation and the existing 4R
framework for DTs, we can now proceed with a direct comparison of both technolo-
gies. In the following discussion, we will explore the similarities and differences
between simulation and DTs.
. Digital Representation: Both simulation and DTs involve creating a virtual repre-
  sentation of a physical system by gathering and understanding data from the
  physical world. The key difference is that a simulation can vary from being very
  realistic to highly abstracted, whereas DT fully describes an existing physical
  element or one that will exist in the physical world [26].
. Real-Time: Simulations may rely on data-driven approaches, but they usually
  lack real-time data integration. Even at the most advanced level of simulation
  capability (S4), real-time data flows are typically absent. Simulation typically
  relies on historical data and performing input analysis of the data to use proba-
  bility distributions or rules to represent that system over time, whereas the DT is
  continuously updated to mirror the real-world counterpart in real time. The real-
  time capability of a DT happens during the R2 and R3 phase; however, the key
  parameters and data pipeline of the DT happens during the first level (R1), which
  requires the establishment of real-time data pipeline integration from sensors, IoT
  devices, or other sources.
. Data Flows: Simulations usually involve one-way data flow, which is established
  at the first capability level (S1). DTs, on the other hand, commonly feature bidi-
  rectional data flow, with the flow from the physical to the virtual domain set up
  during the representation phase (R1), and the virtual to physical data flow estab-
  lished in the final capability level (R4). Bidirectional data flows are not a typical
  component of simulations, even at higher capability levels (S2, S3, and S4).
. Analysis and Visualization: Both simulation and DTs can be used for analysis and
  visualization, which involves using the data from the physical system to analyze
  the system’s behavior and replicate it within a virtual environment. Both can
  replicate identical outcomes when provided with the same inputs as those of the
Contrasting Capabilities: A Comparative Analysis of Simulation …                      45
5 Conclusion
References
 1. Hyre A, Harris G, Osho J, Pantelidakis M, Mykoniatis K, Liu J (2022) Digital twins: Repre-
    sentation, Replication, Reality, and Relational (4Rs). Manuf Lett 31:20–23. https://doi.org/10.
    1016/j.mfglet.2021.12.004
 2. Osho J, Hyre A, Pantelidakis M, Ledford A, Harris G, Liu J et al (2022) Four Rs frame-
    work for the development of a digital twin: the implementation of representation with a FDM
    manufacturing machine. J Manuf Syst 63:370–380. https://doi.org/10.1016/j.jmsy.2022.04.014
 3. Liu M, Fang S, Dong H, Xu C (2021) Review of digital twin about concepts, technologies, and
    industrial applications. J Manuf Syst 58:346–361. https://doi.org/10.1016/j.jmsy.2020.06.017
 4. Ludwig S, Karrenbauer M, Fellan A, Schotten HD, Buhr H, Seetaraman S et al (2018) A 5G
    architecture for the factory of the future 2018
 5. Syafrudin M, Alfian G, Fitriyani NL, Rhee J (2018) Performance analysis of IoT-based sensor,
    big data processing, and machine learning model for real-time monitoring system in automotive
    manufacturing. Sensors (Switzerland) 18. https://doi.org/10.3390/s18092946
 6. Tao F, Zhang H, Liu A, Nee AYC (2019) Digital twin in industry: state-of-the-art. IEEE Trans
    Industr Inform 15:2405–2415. https://doi.org/10.1109/TII.2018.2873186
 7. Melesse TY, Di Pasquale V, Riemma S (2021) Digital Twin models in industrial operations:
    state-of-the-art and future research directions. IET Collaborative Intell Manuf 3:37–47
 8. VanDerHorn E, Mahadevan S (2021) Digital Twin: generalization, characterization and
    implementation. Decis Support Syst 145:113524. https://doi.org/10.1016/j.dss.2021.113524
 9. Walton RB, Ciarallo FW (2022) A framework to manage systems based on digital twin. In: IIE
    annual conference. Proceedings, Institute of Industrial and Systems Engineers (IISE), pp 1–6
10. Jones D, Snider C, Nassehi A, Yon J, Hicks B (2020) Characterising the digital twin: a system-
    atic literature review. CIRP J Manuf Sci Technol 29:36–52. https://doi.org/10.1016/j.cirpj.2020.
    02.002
11. Boschert S, Heinrich C, Rosen R (2018) Next generation digital twin. In: Proceedings of TMCE,
    Las Palmas de Gran Canaria, Spain, pp 7–11
12. Grieves M, Vickers J (2017) Digital Twin: mitigating unpredictable, undesirable emergent
    behavior in complex systems. In: Kahlen F-J, Flumerfelt S, Alves A (eds) Transdisci-
    plinary perspectives on complex systems: new findings and approaches. Springer International
    Publishing, Cham, pp 85–113. https://doi.org/10.1007/978-3-319-38756-7_4
13. Abramovici M, Göbel JC, Savarino P (2016) Virtual twins as integrative components of smart
    products. In: Harik R, Rivest L, Bernard A, Eynard B, Bouras A (eds) Product lifecycle manage-
    ment for digital transformation of industries. Springer International Publishing, Cham, pp
    217–226
14. Schroeder GN, Steinmetz C, Pereira CE, Espindola DB (2016) Digital Twin data modeling with
    automation ML and a communication methodology for data exchange. IFAC-PapersOnLine
    49:12–17. https://doi.org/10.1016/j.ifacol.2016.11.115
15. Shao G, Kibira D (2018) Digital manufacturing: requirements and challenges for implementing
    digital surrogates. In: 2018 Winter Simulation Conference (WSC), IEEE, pp 1226–1237
16. Tao F, Zhang M (2017) Digital Twin shop-floor: a new shop-floor paradigm towards smart
    manufacturing. IEEE Access 5:20418–20427. https://doi.org/10.1109/ACCESS.2017.2756069
17. Gabor T, Belzner L, Kiermeier M, Beck MT, Neitz A (2016) A simulation-based architecture
    for smart cyber-physical systems. IEEE Int Conf Autonomic Comput (ICAC) 2016:374–379
18. Rosen R, von Wichert G, Lo G, Bettenhausen KD (2015) About the importance of autonomy
    and digital twins for the future of manufacturing. IFAC-PapersOnLine 48:567–572. https://doi.
    org/10.1016/j.ifacol.2015.06.141
19. Coronado PDU, Lynn R, Louhichi W, Parto M, Wescoat E, Kurfess T (2018) Part data integra-
    tion in the shop floor Digital Twin: mobile and cloud technologies to enable a manufacturing
    execution system. J Manuf Syst 48:25–33
20. Negri E, Fumagalli L, Macchi M (2017) A review of the roles of digital twin in CPS-based
    production systems. Proc Manuf 11:939–948
48                                                                        A. Wooley and D. F. Silva
21. Kritzinger W, Karner M, Traar G, Henjes J, Sihn W (2018) Digital Twin in manufacturing:
    a categorical literature review and classification. IFAC-PapersOnLine 51:1016–1022. https://
    doi.org/10.1016/j.ifacol.2018.08.474
22. Singh M, Fuenmayor E, Hinchy EP, Qiao Y, Murray N, Devine D (2021) Digital twin: origin
    to future. Appl Syst Innov 4. https://doi.org/10.3390/asi4020036
23. Juarez MG, Botti VJ, Giret AS (2021) Digital twins: review and challenges. J Comput Inf Sci
    Eng 21. https://doi.org/10.1115/1.4050244
24. Lim KYH, Zheng P, Chen CH (2020) A state-of-the-art survey of Digital Twin: techniques,
    engineering product lifecycle management and business innovation perspectives. J Intell Manuf
    31:1313–1337. https://doi.org/10.1007/s10845-019-01512-w
25. Rathore MM, Shah SA, Shukla D, Bentafat E, Bakiras S (2021) The role of AI, machine
    learning, and big data in digital twinning: a systematic literature review, challenges, and
    opportunities. IEEE Access 9:32030–32052. https://doi.org/10.1109/ACCESS.2021.3060863
26. Grieves MW (2023) Digital Twins: past, present, and future. The Digital Twin. Springer,
    Heidelberg, pp 97–121
27. Barricelli BR, Casiraghi E, Fogli D (2019) A survey on Digital Twin: definitions, characteristics,
    applications, and design implications. IEEE Access 7:167653–167671. https://doi.org/10.1109/
    ACCESS.2019.2953499
28. Fan Y, Yang J, Chen J, Hu P, Wang X, Xu J et al (2021) A digital-twin visualized architecture
    for flexible manufacturing system. J Manuf Syst 60:176–201. https://doi.org/10.1016/j.jmsy.
    2021.05.010
29. Wu C, Zhou Y, Pereia Pessôa MV, Peng Q, Tan R (2021) Conceptual digital twin modeling
    based on an integrated five-dimensional framework and TRIZ function model. J Manuf Syst
    58:79–93. https://doi.org/10.1016/j.jmsy.2020.07.006
30. Zhang H, Qi Q, Tao F (2022) A multi-scale modeling method for digital twin shop-floor. J
    Manuf Syst 62:417–428. https://doi.org/10.1016/j.jmsy.2021.12.011
31. Jones D, Snider C, Nassehi A, Yon J, Hicks B (2020) Characterising the Digital Twin: a
    systematic literature review. CIRP J Manuf Sci Technol 29:36–52. https://doi.org/10.1016/j.
    cirpj.2020.02.002
32. Morgan CB (1984) Discrete-event system simulation
33. Beck A (2008) Simulation: the practice of model development and use. J Simul 2:67. https://
    doi.org/10.1057/palgrave.jos.4250031
34. Yin C, McKay A (2018) Introduction to modeling and simulation techniques
35. Banks J (1998) Principles of simulation. Handbook of simulation: principles, methodology,
    advances, applications, and practice, pp 3–30
36. Law AM, Kelton WD, Kelton WD (2007) Simulation modeling and analysis, vol 3. Mcgraw-
    Hill New York
37. Schluse M, Rossmann J (2016) From simulation to experimentable digital twins: simulation-
    based development and operation of complex technical systems. In: ISSE 2016—2016 inter-
    national symposium on systems engineering—Proceedings Papers, Institute of Electrical and
    Electronics Engineers Inc. https://doi.org/10.1109/SysEng.2016.7753162
38. McHaney R (1991) Computer simulation: a practical perspective. Academic Press
39. MOSIMTEC (2022) 4 Types of simulation models to leverage in your business 2022. https://
    mosimtec.com/types-of-simulation-models/. Accessed 17 Dec 2022
40. Harrison RL (2010) Introduction to monte carlo simulation. In: AIP conference proceedings,
    vol 1204. American Institute of Physics, pp 17–21
41. Davidsson P, Holmgren J, Kyhlbäck H, Mengistu D, Persson M (2006) Applications of
    agent based simulation. In: International workshop on multi-agent systems and agent-based
    simulation. Springer, Heidelberg, pp 15–27
42. Moreira ACM, Silva DF (2021) Applying discrete-event simulation and value stream mapping
    to reduce waste in an automotive engine manufacturing plant. Winter Simul Conf (WSC)
    2021:1–12. https://doi.org/10.1109/WSC52266.2021.9715477
43. Banks J (1999) Introduction to simulation. In: Proceedings of the 31st conference on Winter
    simulation: simulation—a bridge to the future, vol 1, pp 7–13
Contrasting Capabilities: A Comparative Analysis of Simulation …                                 49
44. White KP, Ingalls RG (2015) Introduction to simulation. Winter Simul Conf (WSC) 2015:1741–
    1755. https://doi.org/10.1109/WSC.2015.7408292
45. Ingalls RG (2011) Introduction to simulation. In: Proceedings of the 2011 Winter Simulation
    Conference (WSC), pp 1374–88. https://doi.org/10.1109/WSC.2011.6147858
46. Bennett BS (1995) Simulation fundamentals. Prentice Hall International (UK) Ltd
47. Wooley A (2023) Simulation and Digital Twin: contrasting the capabilities and bridging the
    gap 2023
48. Chung CA (2003) Simulation modeling handbook: a practical approach. CRC Press
49. Sokolowski JA, Banks CM (2011) Principles of modeling and simulation: a multidisciplinary
    approach. John Wiley & Sons
50. Bossel H (2013) Modeling and simulation. Springer, Heidelberg
51. Bandyopadhyay S, Bhattacharya R (2014) Discrete and continuous simulation: theory and
    practice. CRC Press
52. Maria A (1997) Introduction to modeling and simulation. In: Proceedings of the 29th conference
    on Winter simulation, pp 7–13
53. Shannon RE (1998) Introduction to the art and science of simulation. In: 1998 winter simulation
    conference. Proceedings (cat. no. 98ch36274), vol 1, IEEE, pp 7–14
54. Carson JS (2005) Introduction to modeling and simulation. In: Proceedings of the Winter
    simulation conference, IEEE, p 8
55. Goldsman D (2007) Introduction to simulation. In: 2007 Winter simulation conference, IEEE,
    pp 26–37
56. Sanchez PJ (2006) As simple as possible, but no simpler: a gentle introduction to simulation
    modeling. In: Proceedings of the 2006 Winter simulation conference, IEEE, pp 2–10
57. White KP, Ingalls RG (2018) The basics of simulation. In: 2018 Winter Simulation Conference
    (WSC), IEEE, pp 147–61
58. Robinson S, Arbez G, Birta LG, Tolk A, Wagner G (2015) Conceptual modeling: definition,
    purpose and benefits. Winter Simul Conf (WSC) 2015:2812–2826. https://doi.org/10.1109/
    WSC.2015.7408386
59. Smith JS, Sturrock DT (2021) Simio and simulation—modeling, analysis, applications, 6th
    edn
60. Davenport TH, Harris JG, Morison R (2010) Analytics at work: smarter decisions, better results.
    Harvard Business Press
61. Wright L, Davidson S (2020) How to tell the difference between a model and a digital twin.
    Adv Model Simul Eng Sci 7. https://doi.org/10.1186/s40323-020-00147-4
62. Ante L (2021) Digital twin technology for smart manufacturing and industry 4.0: a bibliometric
    analysis of the intellectual structure of the research discourse. Manuf Lett 27:96–102. https://
    doi.org/10.1016/j.mfglet.2021.01.003
63. Wooley A, Silva DF, Bitencourt J (2023) When is a simulation a digital twin? A systematic
    literature review. Manuf Lett 35:940–951
Reinforcement Learning for Digital
Twins
Abstract Digital Twins (DTs) aim to ongoingly replicate complex systems through
data acquisition, simulation, and analysis to monitor, optimize, and/or experiment to
achieve systems’ goals. Typically, systems adapt and evolve during their lifetimes,
which requires updating simulation models and analysis as new data arrives or con-
ditions change. The dynamics of environments under which DTs commonly operate
necessitate using a paradigm that can deal with the uncertainties and unprecedented
scenarios that may arise throughout its operation. Reinforcement Learning (RL) is
a learning paradigm that provides tools to do precisely this. It is concerned with
sequential decision-making in dynamic, uncertain environments. In this work, we
discuss the current and potential role of RL in the context of DTs, motivate its usage
through a concrete case study, and finally discuss the opportunities and challenges.
1 Introduction
D. Francis (B)
Department of Engineering Technology, Technical University of Denmark,
Ballerup, Denmark
e-mail: [email protected]
J. Friederich
University of Southern Denmark, Odense, Denmark
e-mail: [email protected]
A. Uhrmacher
University of Rostock, Rostock, Germany
e-mail: [email protected]
S. Lazarova-Molnar
Karlsruhe Institute of Technology, Germany and University of Southern Denmark,
Odense, Denmark
e-mail: [email protected]
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2024          51
M. Grieves and E. Hua (eds.), Digital Twins, Simulation, and the Metaverse, Simulation
Foundations, Methods and Applications, https://doi.org/10.1007/978-3-031-69107-2_3
52                                                                       D. Francis et al.
2 Reinforcement Learning
States                An agent uses its state to select its next action, forming the reinforcement
                      algorithm’s basis. Agent’s state . St is typically a function of the history of
                      encounters between the agent and its environment, i.e., . St = f (Ht ) with
                     . Ht = O1 , R1 , A1 , . . . , Ot−1 , Rt−1 , and is assumed to be Markov, i.e., that
                      the future is independent of the past, given the present. Often, the
                      environmental state is not fully observable by the agent, which makes it
                      formally a Partially Observable Markov Decision Process (POMDP)
Actions               An agent selects actions to maximize its cumulative reward, .A is the set of
                      actions an agent selects from
Rewards               A reward is a scalar feedback signal that indicates how well the agent is
                      doing at step .t,
Discount factor       The discount factor .γ ∈ [0, 1] influences how much an agent values future
                      rewards
Value function        The action-value function .qπ (s, a) is the expected return starting from
                      state .s, taking action .a, and following policy .π. The action-value function
                      can be decomposed into immediate reward and the discounted return from
                      the successor state,
                     .qπ (s, a) = Eπ [Rt+1 + γ qπ (St+1 , At+1 )|St = s, At = a]
fully balanced. Both exploration (trying out new actions to discover their effects)
and exploitation (leveraging known information to maximize rewards) are crucial in
RL. Finding the right balance is challenging, especially in environments with sparse
rewards or deceptive dynamics [4].
Data collection is the DT element concerned with real-time data gathering and storage
from the physical counterpart of the DT. This data is used to generate and update
Reinforcement Learning for Digital Twins                                                       55
DTs. In the design of DTs, the interest has been centered on the technologies for data
collection, such as Internet of Things (IoT), sensors, etc. [7], but in a DT, there are
other aspects of data collection such as how much and when to collect (frequency)
that are also important.
    In order to answer the question of how much data to collect, there are some pre-
vious works that have used pre-training in RL to improve the performance in down-
stream learning tasks while learning the most useful features [8]. It has been found
that when the data is collected passively in finite state episodic Markov Decision Pro-
cesses (MDP), to obtain an .∈-optimal policy, one would need an exponential number
of samples. More formally, to obtain an.∈-optimal policy.Ω(Amin(S−1,H ) ∈ −2 ) episodes
would be needed [9]. The question of when to collect the data can be explored by
looking at an example of RL being employed in the data collection is described in
Table 2. RL has been used in developing optimal data collection strategies in DTs for
wireless communication systems with notable benefits [6]. Such strategies inform
when and what data to collect to refine the RL models.
    Their RL-based data collection method was evaluated against other data collection
schemes in an experimental study, and it was found that their method performed
as well as an oracle-aided baseline method as the number of rounds (more data
samples) increased. Gleeson et al. [10] explored the possibility of reducing the data
collection time as well as computation time by utilizing the parallelization offered
by GPUs. They demonstrate a 13.4x faster implementation of a deep RL framework.
Reducing the state space has been done by including sensor selection as part of the
learning task [11]. Long periods of training and/or data collection are often needed to
obtain success in real-world learning problems [12]. Inspired by the success of data
augmentation in supervised learning, [13] proposed using data augmentation for RL
for pixel-based (image) and state-based data. Their results demonstrated that simple
RL methods can outperform state-of-the-art on several tasks. This approach could
possibly benefit RL for DTs in that the training/data collection time can be reduced
while providing performance boosts.
56                                                                               D. Francis et al.
The data validation of a digital twin consists of tasks, such as imputation, anomaly
detection, etc., which can be solved by supervised and unsupervised learning, respec-
tively [16]. RL has also been used for anomaly detection in sequential and time series
data [17, 18]. In addition, RL can also be used to determine the value of data [15].
This method determines whether the data in hand is useful from a learning/modeling
perspective. Their results indicate that an RL-based approach performed better than
other approaches. Determining the value of data has also been used effectively in
selecting training samples that are similar to the data in a given validation set, which
is quite beneficial in the realm of domain adaptation [15]. The details are shown in
Table 3.
    Often, the collected data contain missing values. A strategy to combat this issue is
to perform data imputation. Data imputation has been approached from unsupervised
and supervised perspectives [20]. RL-based approaches for imputing data have also
been studied with some success [21, 22]. Data measured through sensors have mea-
surement noise. Handling noisy data appropriately is a crucial task. [23] proposed
a noise-robust RL method that demonstrated reliable control in a network control
system. Their method consists of an extended Kalman filter with RL.
In the DT framework outlined in Fig. 1, the model development part is where the
simulation model is generated from the data and extracted knowledge. The DT is
based on this underlying simulation model of the physical counterpart, reflecting its
dynamics and functionalities. The choice of the modeling formalism depends on the
application domain and the goal(s) of the DT. The model’s parameters are updated
at pre-defined intervals to ensure high fidelity. The dynamic nature of the DT makes
adopting an RL model beneficial in this phase. Deployment of the model involves
the model providing insights for decision-making from the PT’s perspective.
Various simulation methods are used in developing, analyzing, and deploying the
digital twin. To make the possible role of RL for selecting and configuring suitable
methods for a DT more concrete, let us look at the selection and configuration of
simulation algorithms. The efficiency of model execution determines how thoroughly
a model can be analyzed and validated. In addition, as the DT is used to inform its
physical twin, the execution efficiency is also decisive for this information to reach
the physical twin in time. Also, the DT needs to reflect changes within its physical
counterpart and shall be functional over a longer period of time; therefore, adaptations
appear as an intrinsic ingredient of the DT. Depending on the simulation model and
the available hardware structure, the efficiency of simulation algorithms varies [24],
implying that for the DT to be effective, simulation algorithms must be selected
and configured automatically. Various approaches have been developed to select and
configure simulation algorithms automatically [25, 26].
    The problem of selecting among the different simulators and their configurations,
given the model and the available hardware, can be posed as a learning problem, either
as a supervised learning problem [27] or as a reinforcement learning problem. To
interpret it as a reinforcement learning problem, i.e., choosing iteratively between
different options with previously unknown rewards, we assume that we can use
part of the simulation, e.g., either some part of a long simulation run or a few of
the replications, to identify a setup that performs best with respect to a user-defined
criterion, e.g., simulation speed. The approach to tackle algorithm selection problems
[28] with reinforcement learning is not unusual [29], even to apply it to reinforcement
learning as a meta learner [30]. To illustrate the challenges, we will shortly sketch
the approach developed in [31], which aims at and analyses automatic, frequent
58                                                                               D. Francis et al.
Table 5 Reinforcement learning to select the most efficient simulator configurations online [31]
 Concept           Description
 Agents                Adaptive simulator
 Environment           Execution environment, simulation model
 States                Model information, e.g., variable values or coupling scheme of
                       components, information about the simulator’s state, e.g., event queue
                       length, and information about the hardware, e.g., available cores
 Actions               Set of simulation algorithms (and components that might be configured)
 Rewards               Rate of event computations per second
 Discount factor       0
 Value function        Q-learning, . Q[s, a] :=
                                                                    ' '
                      . Q[s, a] + α(N [s, a]) · (r + γ · max a ' Q[s , a ] − Q[s, a])
 Model                 model-free
 Policy                f(s,Q,N)
data structures with longer warm-up phases. With .γ = 0, the reinforcement learning
problem reduces to searching (i.e., it is assumed that actions do not affect future
situations and rewards). To serve as a basis of reinforcement learning, this complex,
highly-dimensional information needs to be suitably aggregated into more simple
states. Therefore, dynamic, flexible approaches are required [34], which might ben-
efit from applying learning methods [35]. In any case, a suitable partitioning and
aggregation of the state space is essential for an effective reinforcement learning of
simulator selection and configuration [31].
    Also, the question of when to trigger a reinforcement learning step needs to be
answered: it could depend simply on the time passed since the last learning step or
be triggered by drastic changes in the dynamics. To identify such changes, informa-
tion about the current and previous states of the model, simulator, or hardware and
the current performance must be interpreted to identify suitable points for adapt-
ing (respectively when invoking the reinforcement step) [36]. Generally, change
point detection [37] is an important ingredient to support decisions when dealing
with dynamic systems. Due to the frequent feedback (performance data) between
the agent (adaptive simulator) and environment (execution environment including
model, chosen simulator configuration, hardware), suitable simulator configurations
can be learned effectively [31] and facilitate leveraging algorithms and data structures
tailored to specific requirements and application contexts.
    One could apply reinforcement learning to support other tasks in developing the
DT by method selection and configuration. For example, reinforcement learning
has also been proposed to learn optimization algorithms [38, 39]. The approach
uses guided policy search at the meta-level to train the optimizers and showed good
performance in optimizing neural networks.
    The agent and environment must frequently interact for reinforcement learning
to be effective. Therefore, sample efficiency is critical for the in-situ learning of
methods in the context of DT. Other conditions are that an immediate reward can be
easily calculated and is due to the action (credit assignment problem). Also, it should
be noted that reinforcement learning is only one possibility to solve the algorithm
selection problem [40], even in the context of machine learning approaches, and thus
to support developing or deploying the DT effectively.
AGV 1 Cell 1
solely on the knowledge extraction, model development, and model validation phase
in this case study.
    The following subsections describe the case study system, followed by the DT
extraction procedure. We then examine how an extracted DT model is validated.
Finally, we reflect and discuss the different application areas of RL for this case
study.
Figure 2 provides an overview of the case study system, which is a flow produc-
tion line commonly found in manufacturing systems. The production line is fully
automated and consists of five resource components: a manufacturing execution sys-
tem (MES), two automated guided vehicles (AGVs), and two assembly cells. Both
assembly cells work concurrently, performing the same assembly operation. The
MES controls the production process by initiating new production orders, routing
them to either assembly cell 1 or 2, and marking orders as completed. When a new
production order is initiated and assigned to one of the assembly cells, the AGVs
transport the raw material to the designated cell.
   The AGVs and the assembly cells are susceptible to failures, while the MES is
always fully operational. In the event of a production resource failure, the resource
stops operating, and a repair crew is dispatched to repair the malfunctioning resource.
The maintenance policy in place is purely reactive. The AGV has an unlimited buffer
and a capacity of one, while both assembly cells 1 and 2 have finite buffers and
capacities of one.
   We assume that several types of data can be collected from the case study system,
including event logs, which capture production processes, and state logs, which
record operational state changes of production resources. These two log types are
necessary for the DT extraction process, described in the next section.
Reinforcement Learning for Digital Twins                                                 61
4.2 Digital Twin Model Extraction for the Case Study System
. S P N = (P, T, A, m 0 )
where:
• . P = {P1 , P2 , .., Pp } is the set of places, drawn as circles,
• .T = {T1 , T2 , .., Tq } is the set of transitions along with their distribution functions
   or weights, drawn as bars,
• . A = A I ∪ A O ∪ A H is the set of arcs, where . A O is the set of output arcs, . A I is
   the set of input arcs, and . A H is the set of inhibitor arcs, and each of the arcs has a
   multiplicity assigned to it,
• .m 0 is the initial marking of the Petri net.
    Each transition .Ti can be either timed or immediate. A timed transition is drawn
as a hollow bar and is assigned a probability distribution function that describes
the firing time of the corresponding activity. An immediate transition is drawn as
a filled-out bar and is assigned a weight that describes the firing probability of the
associated event.
    To extract the reliability-focused DT, we first extract a manufacturing process
model and then integrate it with fault models for production resources.
    The extraction of the manufacturing process model utilizes the event log and
involves the following four steps:
1. Identification of a Petri net of the material flow within a production line using
   process discovery algorithms. The material flow is the path that production orders
   follow through the system [41].
2. Determination of transition types (i.e., timed or immediate). Timed transitions
   correspond to the arrival of new production orders or to resource activities.
3. Estimation of probability distributions for timed transitions and extraction of
   weights for immediate transitions.
4. Extraction of resource capacities and buffers.
62                                                                                                                                     D. Francis et al.
                                                                                                                             cell1                      cell1 OK
                                                                                                                            failed
                                                                                                                                           fail
                                                                agv2 failed                                                           weibull(20,90)
                                     fail                                      repair                                                          order completed
                                weibull(20,100)                           lognormal(5,0.2)                                                           mes
                                                                                                                                                      1
  new_order           agv1 OK                     agv1 failed                                   agv2 OK
     mes
exponential(0.3)                     repair                                        fail
                                lognormal(5,0.2)                              weibull(20,100)
                                                            5                                                                             repair
                                                                                                                                     lognormal(3,0.2)
   The extraction of the fault models for production resources utilizes the information
captured by the state log and involves the following two steps:
   Finally, the fault models are integrated into the manufacturing process model
using inhibitor arcs, preventing the corresponding resource activity timed transition
from firing.
   Figure 3 depicts the extracted DT obtained through the described model extraction
process. The figure shows the extracted and parameterized manufacturing process
model as well as the resource fault models. For each timed transition, the corre-
sponding distribution function, including parameters, is shown. For each immediate
transition, the corresponding weight is displayed. Furthermore, the capacity of one
for both assembly cells, as well as their finite buffer sizes, have also been extracted.
Validation is an essential step to ensure that the extracted DT accurately mimics the
real-world system. In [42], we propose to validate extracted models in two phases:
validation of initial model and validation of model at run-time. In the first phase, the
validity of a newly extracted DT model is evaluated to ensure that the model is safe
to deploy in a production environment to support decisions. In the second phase, the
deployed DT model is validated to ensure continuous validity.
Reinforcement Learning for Digital Twins                                                        63
In the following, we investigate and illustrate the application of RL to the case study,
considering all phases of the DT life cycle. Furthermore, we highlight how the case
study system can be optimized using RL and the extracted reliability-focused DT.
Data Collection
The data collection process for reliability-focused DTs can be supported by using
RL to determine the right data streams and the amounts of data needed for model
extraction (Table 6). By adjusting data streams, their size, and frequency of collec-
tion, the impact on the model validation results of the extracted PN model can be
observed and fed back to the agent as a reward. The main challenge in this phase is
the possible large size of the state space, which can lead to high time complexity [43].
Data Validation
RL can be used to to learn and identify patterns, anomalies, or errors within datasets
such as event logs and state logs used for model extraction (Table 7). To do so, spe-
cific metrics or criteria that define the quality and accuracy of the dataset need to be
established. These could include measures such as data consistency, completeness,
accuracy, and adherence to defined standards. The data value determination method
of [15] could serve as a possible solution to selecting useful or error-free data.
Model Extraction
One possible application of RL in the DT model extraction phase could be for material
flow identification (Table 8). RL can be used to interactively extract the material
flow model using process discovery algorithms where the system learns from user
feedback. For instance, when a user corrects or validates certain parts of the material
flow model, the RL system can learn which types of structures are preferred or which
common mistakes to avoid in the following process discovery iterations.
64                                                                                 D. Francis et al.
Model Validation
System optimization
RL can be used in several ways to optimize the real system by making adjustment
to the DT and feeding the gained knowledge back to the real system (Table 10). This
Reinforcement Learning for Digital Twins                                             65
could eventually lead to automated decision support in the real system, unleashing
the full potential of RL. Some of the activities RL could help optimize the real system
are:
• Find optimal parameters to increase/improve a KPI (e.g., production volume,
  resource downtime).
• Use an RL agent to determine routing decisions instead of hard-coded routing
  probabilities (i.e., direct_to_line_1 and direct_to_line_2 transitions).
• Order scheduling: Use an RL agent to determine production order schedule, instead
  of using manually defined order schedules (i.e., new_order transition).
• Buffer size allocation: Use an RL agent to dynamically adjust buffer sizes based
  on system state (i.e., buffer for cell 1 and cell 2).
• Optimal repair scheduling: Use RL agent to schedule and dispatch repair crews.
A deep Q-network model was used in the regulation of a subsystem of the Fermilab
booster accelerator complex with notable success in a test scenario [45][46]. RL has
also been demonstrated to be effective model optimization in a sheet metal assembly
setup [47].
In our study, we investigated how Digital Twins can be enhanced using Reinforcement
Learning at the different steps of Digital Twins’ lifecycle. Reinforcement Learning
is suited to the scenarios found in Digital Twins where sequential decision-making
66                                                                                D. Francis et al.
References
 7. Liu Mengnan, Fang Shuiliang, Dong Huiyue, Cunzhi Xu (2021) Review of digital twin about
    concepts, technologies, and industrial applications. J Manufact Syst 58:346–361
 8. Ghosh D, Bhateja CA, Levine S (2023) Reinforcement learning from passive data via latent
    intentions. Int Conf Mach Learn 11321–11339. PMLR
 9. Xiao C, Lee I, Dai B, Schuurmans D, Szepesvari C (2022) The curse of passive data collec-
    tion in batch reinforcement learning. In: International conference on artificial intelligence and
    statistics. PMLR, pp 8413–8438
10. Gleeson J, Snider D, Yang Y, Gabel M, de Lara E, Pekhimenko G (2022) Optimizing data
    collection in deep reinforcement learning. arXiv preprint arXiv:2207.07736
11. Kishima Yasutaka, Kurashige Kentarou (2013) Reduction of state space in reinforcement learn-
    ing by sensor selection. Artif Life Rob 18:7–14
12. Kalashnikov D, Irpan A, Pastor P, Ibarz J, Herzog A, Jang E, Quillen D, Holly E, Kalakrishnan
    M, Vanhoucke V et al (2018) Qt-opt: Scalable deep reinforcement learning for vision-based
    robotic manipulation. arXiv preprint arXiv:1806.10293
13. Laskin Misha, Lee Kimin, Stooke Adam, Pinto Lerrel, Abbeel Pieter, Srinivas Aravind (2020)
    Reinforcement learning with augmented data. Adv Neural Inf Process Syst 33:19884–19895
14. Williams RJ (1992) Simple statistical gradient-following algorithms for connectionist rein-
    forcement learning. Mach Learn 8:229–256
15. Yoon J, Arik S, Pfister T (2020) Data valuation using reinforcement learning. Int Conf Mach
    Learn 10842–10851. PMLR
16. Chandola Varun, Banerjee Arindam, Kumar Vipin (2009) Anomaly detection: a survey. ACM
    Comput Surv (CSUR) 41(3):1–58
17. Oh M-H, Iyengar G (2019) Sequential anomaly detection using inverse reinforcement learning.
    In: Proceedings of the 25th ACM SIGKDD international conference on knowledge discovery
    and data mining, pp 1480–1490
18. Mengran Yu, Sun Shiliang (2020) Policy-based reinforcement learning for time series anomaly
    detection. Eng Appl Artif Intell 95:103919
19. Fraternali F, Balaji B, Sengupta D, Hong D, Gupta RK (2020) Ember: energy management of
    batteryless event detection sensors with deep reinforcement learning. In: Proceedings of the
    18th conference on embedded networked sensor systems, pp 503–516
20. Emmanuel Tlamelo, Maupong Thabiso, Mpoeleng Dimane, Semong Thabo, Mphago Banyat-
    sang, Tabona Oteng (2021) A survey on missing data in machine learning. J Big Data 8(1):1–37
21. Mei H, Li J, Shi B, Wei H (2023) Reinforcement learning approaches for traffic signal control
    under missing data. arXiv preprint arXiv:2304.10722
22. Yamaguchi N, Fukuda O, Okumura H (2020) Model-based reinforcement learning with miss-
    ing data. In: 2020 eighth international symposium on computing and networking workshops
    (CANDARW). IEEE, pp 168–171
23. Bui V-P, Raj Pandey S, Ana PM, Popovski P (2023) Value-based reinforcement learning for
    digital twins in cloud computing. arXiv preprint arXiv:2311.15985
24. Jeschke M, Ewald R (2008) Large-scale design space exploration of ssa. In: Computational
    methods in systems biology: 6th international conference CMSB 2008, Rostock, Germany,
    October 12-15, 2008. Springer, Berlin, pp 211–230
25. Ewald R (2012) Automatic algorithm selection for complex simulation problems.
    Vieweg+Teubner
26. Gratl FA, Seckler S, Tchipev N, Bungartz H-J, Neumann P (2019) Autopas: Auto-tuning for
    particle simulations. In 2019 IEEE international parallel and distributed processing symposium
    workshops (IPDPSW). IEEE, pp 748–757
27. Ewald R, Himmelspach J, Uhrmacher AM (2008) An algorithm selection approach for simu-
    lation systems. In: 2008 22nd workshop on principles of advanced and distributed simulation.
    IEEE, pp 91–98
28. Rice JR (1976) The algorithm selection problem. In: Advances in computers, Vol. 15, pp
    65–118. Elsevier
29. Lagoudakis MG, Littman ML et al (2000) Algorithm selection using reinforcement learning.
    In: ICML, pp 511–518
68                                                                                 D. Francis et al.
30. Afshar RR, Zhang Y, Vanschoren J, Kaymak U (2022) Automated reinforcement learning: An
    overview. arXiv preprint arXiv:2201.05000
31. Helms T, Ewald R, Rybacki S, Uhrmacher AM (2015a) Automatic runtime adaptation for
    component-based simulation algorithms. ACM Trans Model Comput Simul (TOMACS)
    26(1):1–24
32. Zeigler BP, Praehofer H, Kim TG (2000) Theory of modeling and simulation. Academic Press
33. Himmelspach J, Uhrmacher AM (2007) Plug’n simulate. In: 40th annual simulation symposium
    (ANSS’07). IEEE, pp 137–143
34. Helms T, Mentel S, Uhrmacher A (2016) Dynamic state space partitioning for adaptive sim-
    ulation algorithms. In: Proceedings of the 9th EAI international conference on performance
    evaluation methodologies and tools, pp 149–152
35. Asadi K, Abel D, Littman ML (2020) Learning state abstractions for transfer in continuous
    control. arXiv preprint arXiv:2002.05518
36. Helms T, Reinhardt O, Uhrmacher AM (2015b) Bayesian changepoint detection for generic
    adaptive simulation algorithms. In: Proceedings of the 48th annual simulation symposium, pp
    62–69
37. Aminikhanghahi S, Cook DJ (2017) A survey of methods for time series change point detection.
    Knowl Inf Syst 51(2):339–367
38. Li K, Malik J (2016) Learning to optimize. arXiv preprint arXiv:1606.01885
39. Chen T, Chen X, Chen W, Heaton H, Liu J, Wang Z, Yin W (2022) Learning to optimize:
    a primer and a benchmark. J Mach Learn Res 23(189):1–59. http://jmlr.org/papers/v23/21-
    0308.html
40. Kerschke P, Hoos HH, Neumann F, Trautmann H (2019) Automated algorithm selection: Survey
    and perspectives. Evol Comput 27(1):3–45
41. Lugaresi G, Matta A (2021) Automated manufacturing system discovery and digital twin
    generation. J Manufac Syst 59:51–66, Apr 2021. ISSN 0278-612 https://doi.org/10.1016/j.
    jmsy.2021.01.005. https://www.sciencedirect.com/science/article/pii/S0278612521000054
42. Friederich J, Lazarova-Molnar S (2023) A framework for validating data-driven discrete-event
    simulation models of cyber-physical production systems. In: Proceedings of the 2023 winter
    simulation conference. San Antonio, USA, IEEE
43. Dietterich TG (2000) An overview of maxq hierarchical reinforcement learning. In: Inter-
    national symposium on abstraction, reformulation, and approximation. Springer, Berlin, pp
    26–44
44. Schena L, Marques P, Poletti R, Ahizi S, Van den Berghe J, Mendez MA (2023) Reinforce-
    ment twinning: from digital twins to model-based reinforcement learning. arXiv preprint
    arXiv:2311.03628
45. Kafkes D, Schram M (2021) Developing robust digital twins and reinforcement learning for
    accelerator control systems at the fermilab booster. arXiv preprint arXiv:2105.12847
46. John JS, Herwig C, Kafkes D, Mitrevski J, Pellico WA, Perdue GN, Quintero-Parra A, Schup-
    bach BA, Seiya K, Tran N et al (2021) Real-time artificial intelligence for accelerator control:
    A study at the fermilab booster. Phys Rev Accelerators Beams 24(10):104601
47. Cronrath C, Aderiani AR, Lennartson B (2019) Enhancing digital twins through reinforcement
    learning. In: 2019 IEEE 15th international conference on automation science and engineering
    (CASE). IEEE, pp 293–298
48. Yang L, Wang M (2020) Reinforcement learning in feature space: matrix bandit, kernels, and
    regret bound. In: International conference on machine learning. PMLR, pp 10746–10756
49. Osband I, Van Roy B (2014) Near-optimal reinforcement learning in factored mdps. Adv Neural
    Inf Process Syst 27
A Framework for the Credibility
Evaluation of Digital Twins
Han Lu, Lin Zhang, Kunyu Wang, Zejun Huang, Hongbo Cheng,
and Jin Cui
Abstract The widespread applications of Digital Twins are facing a significant chal-
lenge due to the lack of systematic and effective credibility evaluation methods. To
address this issue, this chapter analyzes the dynamic evolution, virtual-real interac-
tivity, and other key features of Digital Twins. Subsequently, a credibility evaluation
framework for Digital Twins is proposed, encompassing the concept of Digital Twins
credibility, a multi-dimensional and multi-level credibility evaluation index system,
and a credibility evaluation methodology. The evaluation process is elucidated using
the robotic arm Digital Twin as an example, thereby providing valuable guidance for
the evaluation and construction of Digital Twins.
1 Introduction
A Digital Twin can be taken as a digital model of a physical object that under-
goes real-time self-evolution through data integration from its physical counterpart,
ensuring consistent synchronization throughout its entire lifecycle [1]. Simulations
based on the model can help to optimize and make decisions on the physical object
[1, 2]. The physical object can be any type of physical system, such as a transporta-
tion system, a factory, a production line, or an equipment. The Digital Twins in
this chapter mainly focus on equipment. The equipment can be the product of the
manufacturing industry or the equipment used in the manufacturing process, e.g.,
the aircraft, car, machine tool, or robot. The adoption of Digital Twins in the equip-
ment’s lifecycle offers numerous benefits, including enhanced equipment innovation
and design, reduced development cycles, cost reductions, and improved autonomy in
equipment operation and maintenance [3–8]. During the research and design stage
of equipment development, the utilization of modeling and simulation-based system
engineering techniques enables activities to be conducted within the digital realm [9,
10]. Consequently, design schemes can be thoroughly tested and pre-demonstrated,
leading to significant reductions in physical experiment investments.
    Credibility is a metric to measure the “correctness” of a model and the model’s
results for a specific use or purpose [11], which is the degree of user’s confidence in the
correctness of a model [8]. The concept of Verification, Validation, and Accreditation
(VV&A) was proposed by the US Department of Defense (DoD) [12] to guarantee
the credibility of modeling and simulation [13]. How to evaluate the credibility of a
model is a critical task and foundation of VV&A. As a self-evolving and dynamic
model, Digital Twin’s credibility is much more difficult to be evaluated. In practice,
lacking of credibility evaluation for Digital Twins is one of the critical bottlenecks
in the widespread utilization of Digital Twins, as well as Digital Twins. Only a
credible Digital Twin can effectively capture the equipment’s characteristics and
status in a timely manner, thereby assisting decision-makers in making accurate
judgments. Considering that the Digital Twin maintains real-time data interaction
with its physical counterpart, the usage of an untrustworthy Digital Twin may lead
to failed tasks and even catastrophic consequences.
    The research on credibility evaluation of Digital Twins is currently limited world-
wide. Consequently, the widespread adoption and extensive application of Digital
Twins face significant barriers. Existing credibility evaluation methods for Digital
Twins primarily focus on visual resemblance and simulation output results, which
are insufficiently comprehensive and scientific for complex Digital Twins.
    Firstly, it is essential to recognize that a digital model cannot perfectly repli-
cate a physical object. The excessive pursuit of realism or high accuracy may result
in unnecessary complexity and introduce additional uncertainty, thereby reducing
model credibility and usability. It is crucial to strike a balance between accuracy
and complexity to facilitate efficient model computation and utilization. Secondly,
evaluating credibility solely based on the consistency of output data collected at a
certain time interval between the Digital Twin and its physical counterpart is insuf-
ficient. While external performance similarity is valuable, it does not guarantee that
the internal mechanistic structure of the Digital Twin matches that of the physical
object. Thirdly, relying on data samples from a specific time period fails to represent
all the potential output variations of the model. Therefore, the credibility of Digital
Twins cannot be solely determined by the visual resemblance or simulation output
results.
    In summary, the credibility evaluation of Digital Twins demands a comprehensive
and systematic approach. There is an urgent need for the development of a compre-
hensive evaluation theory and methodology. Such methodologies can offer practical
solutions for assessing the credibility of Digital Twins. Moreover, they can provide
guidance on constructing credible Digital Twins, thus ensuring their credibility for
various applications.
A Framework for the Credibility Evaluation of Digital Twins                            71
2 Related Work
During the past decade of concept development related to DTs, most of the litera-
ture focused on discussing the concept itself and its underlying connotations [14–
17]. However, there remains a notable dearth of a systematic credibility evaluation
framework for reference within the field. Evaluating a DT entails assessing both the
DT model itself and the real-time data utilized for its construction and evolution.
While theoretically, data changes should be incorporated and reflected in the model,
a separate evaluation of the credibility of the data can help prevent the introduc-
tion of untrustworthy data into the DT, thereby enhancing DT credibility. Despite
the limited resources specifically addressing credibility evaluation for DTs, there is
considerable research available in the field of M&S that may serve as a valuable
resource to develop credibility evaluation methods for DTs.
The credibility of data used in a DT significantly influences the credibility of the DT.
These data encompass a large number of sensor data and non-sensor data related to
human activities.
   In the case of sensor data, the reliability can be compromised if the sensor is
damaged or experiences failure [18]. Additionally, if a sensor fails to function prop-
erly within a sensor network, the data from that sensor node may be considered
unreliable [19]. Even under normal circumstances, sensor data can be influenced
by various factors such as the working environment [20], as well as the installation
distance and angle of the sensors [21]. To evaluate the credibility of sensor data itself,
Liao et al. [22] utilized a cumulative residual Chi-square check to compare current
data with historical data and determine their credibility.
   Digital Twin involves human participation during both the construction and oper-
ation decision-making stages. The credibility of the data in Digital Twin systems has
been evaluated from various perspectives. For instance, the influence of ambiguity
and authority on data credibility has been studied through psychological evaluations
[23]. Additionally, the credibility of human-related propositions has been assessed
based on a priori knowledge and a posteriori knowledge [24].
The existing methods for model credibility evaluation mainly come from the
M&S. The specific methods for assessing model credibility can be categorized
into three groups: qualitative analysis methods, quantitative analysis methods, and
comprehensive analysis methods.
72                                                                             H. Lu et al.
Fig. 1 Credibility of DT
74                                                                            H. Lu et al.
(1) Model life cycle: A Digital Twin is developed around mapping the whole life
    cycle of equipment (design, processing, testing, operation, and maintenance).
    Meanwhile, the Digital Twin model also has a complete life cycle for any life
    cycle stage of equipment, including design, construction, use, evolution, and
    management. Each stage has an independent model form and function focus.
    Therefore, it is necessary to investigate the life cycle stage of the twin model
    and select the appropriate essential credibility metrics and evaluation methods
    according to the user requirements and model characteristics.
(2) Model form: Different life stages correspond to various model forms, such as
    conceptual models, mathematical models, and simulation models. The concep-
    tual model is a formal interpretation of requirements. The mathematical models
    are divided into mechanism models and non-mechanism models. Simulation
    models are the code expression of conceptual models and mathematical models,
A Framework for the Credibility Evaluation of Digital Twins                         75
     the size of basic constituent units within the scope. There is a strong corre-
     lation between spatial scale and granularity. Due to the limitation of evalua-
     tion resources, the smaller the investigation granularity, the smaller the corre-
     sponding spatial scale, and the larger the spatial scale, the larger the investigation
     granularity.
   In addition to the above six dimensions, the complexity of the model structure,
the model composition, the relationship between models, the model interoperability,
and the interaction between models and real-time data should be considered in the
evaluation. Therefore, it is necessary to have a set of index system involving multi-
dimensional elements and corresponding hierarchical evaluation methods.
The real-time evolutionary nature of Digital Twins couples the construction process
and application process of twin models. Unlike the traditional basic model where
there is a clear boundary between construction and application, the twin model keeps
changing along with the changes of physical equipment during its applications. Thus,
the evaluation of the construction and application should be regarded as part of the
whole process, which can be called as a “modeling-application-evaluation cycle”.
As shown in Fig. 3, the key output and state data of the running equipment are
collected in real time. According to the difference between the real and virtual data,
the synchronized twin model identifies the biased components and conducts online
evolutionary calibration. Thus, the consistency between the model and the equip-
ment in terms of mechanism, structure, and parameters could be maintained. Once
the credibility of the twin model meets the requirements, various simulation-based
applications are allowed to be executed on the equipment.
    The effect of the dynamic evolution process is shown in the credible evolution
box in Fig. 3, where the blue line refers to the states and outputs of the physical
equipment and the red line shows those of the twin model. The physical equipment
is influenced by the environment and teams including the equipment suppliers, the DT
users, and the evaluators. Compared with the equipment, the initial twin model still
has difference which will increase with the equipment running. The trigger module
calculates the incoming difference in real time and triggers the evolution process
according to the preset threshold. Then, the evolution module receives the messages
and starts assimilating the data.
    As the number of loop iterations increases, the output error of the twin model will
be gradually reduced and eventually converged with the equipment output. In this
process, the evolution method will be optimally adjusted according to the obtained
credibility. Optimization aspects include the precision, over-advancement, and light
weighting, so that the twin model can continuously achieve higher credibility under
the given resource conditions.
A Framework for the Credibility Evaluation of Digital Twins                       77
The real-time interaction between the equipment and the twin model is the premise
of DT dynamic evolution, which includes forward calibration and feedback opti-
mization. Forward calibration transfers data from the equipment to the Digital Twin,
including real-time collection of equipment operating information, data generated
from various software, manually updated data like text drawings. Feedback optimiza-
tion delivers optimized control signals and scheduling planning from the Digital Twin
to the equipment, to promote the iterative updating of the equipment. As the equip-
ment is in operation, the user’s application requirements for the equipment may
change from time to time, which will deeply affect the credibility connotation of
the relevant Digital Twin. So, it is necessary to recognize the changes in the user’s
requirements and update the method of credibility evaluation in time.
78                                                                               H. Lu et al.
   The two directions of real-time interactions are relatively independent but closely
related. Forward calibration is the foundation of credibility insurance. The quality and
timeliness of the data in this process will deeply affect the credibility of subsequent
predictions and simulations. Feedback optimization is the purpose of DT. If the
performance of the system is degraded in this step, it is likely to lead to a vicious cycle.
So actual feedback should be implemented cautiously with sufficient simulation
demonstration and risk prevention measures. During the evaluation, the independence
and relevance of the two directions should be fully considered from the perspectives
of both data and the twin model.
As shown in Fig. 4, the credibility evaluation of the Digital Twin is mainly composed
of three parts: the construction of the index system, the basic evaluation, and the
evolution evaluation. The Digital Twin is a simulation model constructed around the
requirements. So, the multi-dimensional and multi-level index system required for
the evaluation should be obtained by analyzing the user requirements for the model.
The appropriate evaluation methods are then constructed to form a methodological
framework, by combining the characteristics of the model and investigation results
of the requirements.
   On a given evaluation time window, the twin model is composed of a sequence
of basic models arranged on the time axis. The model evolution which occurs at a
time point generates a basic model for a new time point according to the previous
basic model; thus, an update of the twin model is completed. The basic evaluation
involves every basic model in the model sequence. The first basic evaluation needs
to be conducted in a comprehensive and detailed manner. The subsequent evalua-
tions are more lightweight and could only focus on the changed components. The
evolution evaluation examines the changes in the basic model sequence. The correct-
ness of each basic model and the timeliness of the overall basic model sequence are
influenced by each other. In general, the pursuit of higher correctness would lead to
lower timeliness. The examination of uncertainty gives the confidence distribution
for each quantitative value. The overall combination of correctness, timeliness, and
confidence distribution forms the credibility of the twin model.
    When a closed loop is formed by connecting the basic model with the equipment,
the evaluator should continuously conduct credibility evaluation throughout the life
cycle of the model according to the basic characteristics of each stage. The following
section will focus on the framework of the index system.
Although simulation technology has been developed for more than 70 years, there
is no unified index system about credibility.
    Due to the complexity of the object, the credibility of the twin model needs to be
examined from multiple aspects. Synthesizing the various types of index systems on
credibility in different fields, and considering the characteristics of Digital Twins,
this chapter proposes a framework of credibility evaluation index system for Digital
Twins (Fig. 5). The superscript of the index symbols is the classification to which it
belongs, and the subscripts is the meaning of the indexes. The system consists of four
major parts, which can be further refined and expanded according to the meanings
of the indexes below, combined with the needs and characteristics of the evaluation
object.
Data is the foundation of Digital Twins construction and has a significant impact on
DT credibility. DT data is often characterized by real-time, high concurrency, and
multi-source heterogeneity. Data evaluation examines both the source that generates
the data and the data itself. The indexes are shown in Table 1.
When the basic model has formed a closed loop with the physical equipment, the
model parameters, structure, and mechanism will be updated based on the data
collected from the physical system. Model evolution is an important feature which
separates Digital Twin models from traditional models. As shown in Table 3, the
dynamic indexes are for the evolution process, which mainly examine whether the
virtual model correctly follows the physical object, whether the response to various
evolutionary events is timely, and whether the whole evolution trend converges.
Based on the multi-dimensional correlation analysis of the model in the preparation
of Sect. 5.2.1, the dynamic changes in the model structure and mechanism can be
parameterized for consistency evaluation.
The above three indexes are primary, while the supporting indexes are often used as
references. These include maturity of capabilities and maturity of models [43], which
are more difficult to obtain and has indirect impact on model credibility (Table 4).
5.2.1 Preparation
(1) Analyze the model composition of the evaluation object. According to Fig. 2, a
    multi-dimensional decomposition diagram of the complex model (Fig. 7) should
A Framework for the Credibility Evaluation of Digital Twins                         83
Fig. 7 Multi-dimensional
decomposition diagram of
complex models
(1) Determine the specific index system. The specific indexes are further refined by
    considering both the characteristics of the model and the requirement analysis. A
    complete multi-dimensional hierarchical index system would be formed based
    on the index system shown in Fig. 5. Furthermore, the acceptable value range
    of each index is determined under the guidance of quantitative requirement
    analysis, which would serve as a benchmark for the correctness of the subsequent
    evaluation.
(2) Select appropriate methods. Eighty-one existing credibility evaluation methods
    are sorted out and organized into in eight categories (Table 5). Different indexes
    can be quantified and assessed using the appropriate methods.
   For each single index, it is necessary to select the appropriate method or combi-
nation of methods from the list by considering the life cycle stage where the model
serves, model characteristics, and user requirements. There is still a lack of evalu-
ation methods for some indexes, which is also an important part of future study. In
84                                                                                H. Lu et al.
(3) Evaluation of supporting elements. For the maturity of capabilities and models,
    a comprehensive examination of the process of their evaluations and improve-
    ments under multiple application scenarios and different needs is performed.
    These experiences could serve as references for credibility estimation of unprac-
    ticed scenarios under the same scope, which can be referred to in the literature
    [43].
86                                                                            H. Lu et al.
Evolution evaluation focuses on the evolution process of the simulation model. There
are some similarities between the dynamic characteristic analysis of the traditional
basic model and the evolution evaluation of the twin model. However, they are signif-
icantly different. On the one hand, the evolution of the model makes the dynamic
characteristic curves change continuously. So the evolution evaluation cannot only
analyze the curves at a specific time point, but rather, it has to comprehensively
consider the entire sequence of curves on a period of time window. The key element
to be examined during evolution evaluation is the change process of the basic model
sequence. On the other hand, the twin model shows the property of modeling–appli-
cation–evaluation cycle as mentioned in Sect. 4.2. Thus, during an evolution process,
the object keeps being assessed and the twin model may change accordingly. Then,
the evaluation criteria should be changed as well.
    In general, there are three types of credibility evaluation triggering methods corre-
sponding to the changes of the object: time-driven, event-driven, and hybrid-driven.
There are also four types of time windows, which are shown in Fig. 9. In the scenario
of online dynamic evaluation of Digital Twins, it is necessary to divide the time
windows according to the actual triggering mechanism, periodicity, and overlapping
                                                                     Dy
situation. Generally, the evolution-driven discriminative index Id is evaluated using
                                        Dy
F23, F31, and the sensitivity index Ise is analyzed using methods such as F51. For a
                                                                  Dy
single time window evolution process, its timeliness index It can be checked using
                                                       Dy
methods such as F71, and its convergence index Icv can be analyzed using methods
                                                                                Dy
such as F73. For the corresponding evolution result, the systemic index Isy of the
                                                                                       Dy
evolution can be assessed using methods such as F74, and the correctness index Icv
of the evolutionary result can be analyzed using methods such as F72.
    The dynamic integration method is based on multiple time windows. The intrinsic
correlation between multiple triggering mechanisms is analyzed. Then, the dispersion
of the weights on the multiple time windows are obtained, so that the dynamic index
values could be continuously synthesized from the basic index values.
A Framework for the Credibility Evaluation of Digital Twins                       87
   Uncertainty is examined at the last step. For each index value, the data, evalua-
tion methods, and management methods involved in the acquisition process should
be examined. Both the inherent uncertainty, including systematic parameter errors,
environmental perturbations, and the cognitive uncertainty, including inappropriate
model forms and empirical parameter errors are included. The failure probability of
the assessed value brought about by the uncertainty is then quantified and uniformly
mapped to the distribution of confidence intervals. By combining this distribution
with its degree of correctness and timeliness, the degree to which the model actu-
ally meets the demand, which is the degree of credibility, could be obtained. Based
on the requirement-model mapping, the single-index value, multi-dimensional inte-
gration value, and dynamic evaluation value are transformed into the final dynamic
multi-dimensional distribution of the credibility.
AG95 claw, sensors such as thermometers, hall sensors, voltage and current sensors,
and communication devices such as routers and edge gateways. The software used
is Unity, ANSYS Workbench, C#.
    The credibility evaluation started with a multi-dimensional decomposition of the
robotic arm DT model by the modeling experts. The process is followed by analyzing
the requirements for “grasping a target and placing it in a specified area”. The two
key inspection points are then identified: “stable grasping with appropriate force”
and “timely routing and obstacle avoidance”. Afterward, the model is decomposed
step by step from qualitative to quantitative to obtain the specific index system and
the acceptable range of requirements. In this step, the normalization index InS and
completeness index IcS of the requirements-model mapping are assessed.
    A unidimensional evaluation focusing on multi-disciplinary domains and simu-
lation forms was then conducted. Among the multi-disciplinary domains, geometry,
kinematics, mechanics, control, electricity, and magnetism have a strong influence
on the grasping requirements, so the correctness of their disciplinary models is the
focus of the evaluation. Meanwhile, the correctness of the system structure, code and
interface specification, and logical correctness of the simulation models are the main
investigations of the polymorphic dimension.
    Thereafter, the compatibility index IcD of the acquisition communication device
                            D
and the authority index Iao    of data sources such as vendors and built-in software
of the robotic arm were evaluated using formal review and expert scoring methods.
                           D
The authenticity index Iae    of historical modeling data, real-time status data, and
requirement data was evaluated in depth using a variety of data validation methods.
A Framework for the Credibility Evaluation of Digital Twins                            89
The overall technology maturity was assessed using subjective and formal methods
in conjunction with historical data related to model development and application.
The uncertainty of each evaluation component was analyzed using dataset validation
and result comparison methods.
   After completing the basic evaluation, a multi-disciplinary and multi-dimensional
integration was performed using a modified DT theory of evidence. The integration
was based on the following Dempster synthesis rule:
                                  
                                      0,
                                      
                                         (A = )
                        m(A) =         Ai ∩Bj =A   m1 (Ai )m2 (Bj )                ,
                                                   1−K
                                                                      , (A = )
                                                                           
where K is the conflict factor which meets K = Ai ∩Bj =A m1 (Ai )m2 Bj < 1 and
reflects the degree of conflict between the various pieces of evidence.
    On this basis, the two key evolution processes of the robotic arm attitude change
and real-time deformation of the gripped object are assessed. The process and results
of the calibration of each parameter and the geometrical structure change are system-
atically examined. The time-frequency domain analysis and the convergence anal-
ysis are comprehensively used to obtain the six index values of the whole evolution
process. The quantitative value of the model credibility is given by combining with
the uncertainty analysis.
    In the credibility synthesis process, the weights of each index are given mainly
based on the analysis of scene requirements and experts’ experience. The weights
assigned to the data indexes are relatively small because the experimental environ-
ment is relatively simple and stable, and the robotic arm as well as the related acqui-
sition equipment has high reliability. High weights are assigned to the indexes related
to unidimensional and coupled evaluation due to the high impact of the correctness
of the discipline model on the grasping requirements. Because the real-time calibra-
tion of the robotic arm attitude changes and the deformation of the grasped object
has a greater impact on the grasping and obstacle avoidance needs, high weights are
also assigned to the evolution-related indexes. The other indexes have less influence
on the application requirements of the robotic arm twin model and are uniformly
assigned smaller weights. A multi-objective optimization algorithm will be used in
the subsequent study to determine the optimal weight allocation for each index.
    To validate the credibility evaluation methods, four types of DT models with
different credibility are constructed to form a basic model benchmark, as shown in
Fig. 11. The grasping effects of the four types of DT models are shown in Fig. 12.
    The horizontal axis is the time axis, and the vertical axis is the distance between
the grasped object and the end point. The evaluation calculation results give the
order of model credibility as Model 3 > Model 4 > Model 2 > Model 1, which is
consistent with the experimental performance results. Thus, the effectiveness of the
credibility evaluation framework proposed in this chapter is preliminarily validated.
It is necessary to expand the library of benchmark models in terms of equipment type,
model mechanism, and complexity of scenario requirements, so that the credibility
evaluation framework could be developed and validated in a wider scope.
90                                                                                H. Lu et al.
7 Conclusions
This chapter analyzed three typical characteristics of Digital Twins: dynamic evolu-
tion, real-time interaction, and system complexity. On this basis, the definition of
credibility of Digital Twin is given. Then, a set of credibility evaluation index system
is proposed based on previous researches and practical experiences. The general
process of DT credibility evaluation is described afterward. The credibility evaluation
framework is validated through experiments.
A Framework for the Credibility Evaluation of Digital Twins                                     91
    In the future research, the evaluation index system will be further extended and
refined so that credibility can be described more comprehensively and accurately.
Moreover, some single-index evaluation methods will be improved or supplemented,
such as uncertainty quantification method and credibility integration method. In
addition, more effective quantification methods will be developed for the credibility
evaluation of the evolution model. The refinement and combination of evaluation
methods would be further explored according to the commonality and difference of
the geometric appearance model, the system structure model, and the mechanism
model. The benchmark library of Digital Twin models would be extended to better
validate the generalization capability of the credibility evaluation framework. An
evaluation software system would be developed based on the framework to support
the automatic credibility evaluation to a certain extent.
References
 1. Zhang L, Zhou L, Horn BKP (2021) Building a right digital twin with model engineering. J
    Manuf Syst 59:151–164
 2. Shafto M, Conroy M, Doyle R et al (2010) Technology area 11: modeling, simulation, infor-
    mation technology and processing roadmap. NASA Office of Chief Technologist, November
    2010
 3. Boschert S, Rosen R (2016) Digital Twin-the simulation aspect. Springer International
    Publishing, Cham, pp 59–74
 4. Bielefeldt B, Hochhalter J, Hartl D (2015) Computationally efficient analysis of SMA sensory
    particles embedded in complex aerostructures using a substructure approach. In: ASME 2015
    conference on smart materials, adaptive structures and intelligent systems. ASME, New York,
    USA, V001T02A007
 5. Liu S, Bao J, Lu Y et al (2021) Digital Twin modeling method based on biomimicry for
    machining aerospace components. J Manuf Syst 58(Part B):180–195
 6. Zhang L (2020) Cold thinking about the digital twin and the modeling and simulation techniques
    behind it. J Syst Simul 32(4):1–10
 7. Zhang L, Lu H (2021) Discussing digital twin from of modeling and simulation. J Syst Simul
    33(5):995–1007
 8. Tuegel E J (2012) The airframe digital twin: some challenges to realization. In: 53rd AIAA/
    ASME/ASCE/AHS/ASC structures, structural dynamics and materials conference. 20th AIAA/
    ASME/AHS adaptive structures conference 14th AIAA, 1812
 9. Huang J, Gheorghe A, Handley H et al (2020) Towards digital engineering: the advent of digital
    systems engineering. Int J Syst Syst Eng 10(3):234–261
10. Zhang L, Wang K, Laili Y et al (2022) Modeling & simulation based system of systems
    engineering. J Syst Simul 34(2):179–190
11. Rabeau R (2013) Credibility in modeling and simulation. In: Simulation and modeling of
    systems of systems, pp 99–157
12. Department of Defense (2007) DoD Modeling and Simulation (M&S) Management. Directive
    Number 5000.59, 8 Aug 2007
13. Sargent RG (2015) Model verification and validation. In: Modeling and simulation in the
    systems engineering life cycle: core concepts and accompanying lectures. Springer London,
    pp 57–65
14. Grieves M (2019) Virtually intelligent product systems: digital and physical twins. In: Flumer-
    felt S et al (eds) Complex systems engineering: theory and practice. American Institute of
    Aeronautics and Astronautics, pp 175–200
92                                                                                   H. Lu et al.
15. Lim KYH, Zheng P, Chen CH (2020) A state-of-the-art survey of Digital Twin: techniques,
    engineering product lifecycle management and business innovation perspectives. J Intell Manuf
    31(6):1313–1337
16. Liu M, Fang S, Dong H et al (2021) Review of digital twin about concepts, technologies, and
    industrial applications. J Manuf Syst 58:346–361
17. Lu Y, Liu C, Kevin I et al (2020) Digital Twin-driven smart manufacturing: connotation,
    reference model, applications and research issues. Robot Comput Integr Manuf 61:101837
18. Sun Y, Luo H, Das SK (2012) A trust-based framework for fault-tolerant data aggregation in
    wireless multimedia sensor networks. IEEE Trans Dependable Secure Comput 9(6):785–797
19. Jiang J, Han G, Wang F et al (2014) An efficient distributed trust model for wireless sensor
    networks. IEEE Trans Parallel Distrib Syst 26(5):1228–1237
20. Miao CL (2014) Research on trustworthiness problems in wireless sensor networks. University
    of Science and Technology of China
21. Mao HM (2016) Research of ultrasonic sensor detection model based on credibility evaluation.
    Shenyang Jianzhu University
22. Liao M, Liu J, Meng Z et al (2021) A sins/sar/gps fusion positioning system based on sensor
    credibility evaluations. Remote Sens 13(21):4463
23. Levelt P, Caramazza A (2007) The Oxford handbook of psycholinguistics. Oxford University
    Press, USA
24. Barber A, Robert JS (2010) Concise encyclopedia of philosophy of language and linguistics.
    Elsevier
25. Tang JB (2009) Research on the credibility of warfare simulation system. Graduate School of
    National University of Defense Technology
26. Liao Y, Feng XJ, Zhang J et al (2003) Research on calibration, verification and validation
    (VVA) technology for strategic missile system modeling and simulation. In: Proceedings of
    the 2003 National Academic conference on system simulation, pp 187–192
27. Yang M, Zhang B, Wang ZC et al (1999) VV&A and credibility evaluation of distributed
    interactive simulation system. J Harbin Inst Technol 05:48–51
28. Kim CJ, Lee D, Hur S et al (2017) M&S VV&A for aerospace system development: M&S life-
    cycle model and development paradigm. J Korean Soc Aeronautical Space Sci 45(6):508–516
29. Zhou JH, Xue JJ, Li HY et al (2020) Thinking on digital twin for weapon system. J Syst Simul
    32(04):539–552
30. Zhang SL, Ye MC (2006) Simulation credibility evaluation method for missile weapon system.
    Comput Simul 23:48–51
31. Zhen BT, Hu Y (2015) Research on foreign weapon and equipment modeling and simulation
    credibility assessment methods. Qual Reliab 2:6–9
32. Sim SH, Lee JC (2014) On improving the verification, validation and accreditation process by
    including safety requirements in M&S-based development of weapon systems. J Korea Safety
    Manage Sci 16(4):123–131
33. Beydoun G, Low G, Bogg P (2013) Suitability assessment framework of agent-based software
    architectures. Inf Softw Technol 55(4):673–689
34. Acar E (2015) Effect of error metrics on optimum weight factor selection for ensemble of
    metamodels. Expert Syst Appl 42(5):2703–2709
35. Li N, Dong L, Zhao L et al (2021) A credibility evaluation method for complex simulation
    systems based on interactive network analysis. Simul Model Pract Theory 110:102289
36. Ferson S, Oberkampf WL (2009) Validation of imprecise probability models. Int J Reliab Saf
    3(1–3):3–22
37. Li W, Chen W, Jiang Z et al (2014) New validation metrics for models with multiple correlated
    responses. Reliab Eng Syst Saf 127:1–11
38. Dornheim H, Brazauskas V (2011) Robust–efficient credibility models with heavy-tailed
    claims: a mixed linear models perspective. Insurance Math Econ 48(1):72–84
39. Liang J, Bai Y, Bi C et al (2013) Adaptive routing based on Bayesian network and fuzzy
    decision algorithm in delay-tolerant network. In: 2013 IEEE 10th international conference on
    high performance computing and communications & 2013 IEEE international conference on
    embedded and ubiquitous computing. IEEE, pp 690–697
A Framework for the Credibility Evaluation of Digital Twins                               93
40. Hu XL (2011) Dynamic data driven simulation. SCS M&S Mag 5:16–22
41. Wang K, Zhang L, Jia Z et al (2023) A framework and method for equipment digital twin
    dynamic evolution based on IExATCN. J Intell Manuf 1–13
42. Wang K, Zhang L, Cheng H et al (2023) A lifelong learning method based on event-triggered
    online frozen-EWC transformer encoder for equipment digital twin dynamic evolution. IEEE
    Internet Things J
43. Zhang L, Liu Y, Laili Y et al (2020) Model maturity towards modeling and simulation:
    concepts, index system framework and evaluation method. Int J Modeling Simul Sci Comput
    11(03):2040001
On the Importance of Simulation
and Digital Twin for Industrial
Applications
Abstract The term Digital Twin has become ubiquitous recently as virtual repre-
sentation of a physical asset to leverage business opportunities, i.e., with a dedicated
purpose. It consists of very heterogeneous data and can take very different shapes,
from the collection of datasheet information together with a visual reproduction of
the real object via the compilation of operational data to the application of physics-
based simulations or predictions based on artificial intelligence (AI). Obviously, any
combination thereof is also possible. The applications of Digital Twins are spread
over the whole lifecycle of products and systems including the interaction of Digital
Twins at different hierarchical levels. In addition, Digital Twins can create a connec-
tion between different value chains. An example is the exchange of Digital Twins
along the supply chain with a link to the shop floor, which then enables the execution
of cooperative simulations for different purposes. In our book chapter, we investigate
the role of simulation for a Digital Twin with respect to value chains, the lifecycle
phases, and their business value. Furthermore, in the age of the Industrial Metaverse
the importance of simulation for Digital Twins is getting even more important, as
immersion, interaction, and collaboration in real-time play a greater role, making
it necessary to have fast and good enough predictions of the behavior of industrial
assets. Novel algorithms and machine learning (ML) methods enhance the capabili-
ties of the traditionally engineering-focused simulation tools. The mentioned aspects
will be illustrated with several examples of real-life industrial applications.
1 Introduction
The efficient development and optimal operation of automated systems are well-
known, but still challenging. Automated systems are mostly complex technical
systems that are controlled by (automation) software and increasingly networked with
one another, which transforms technical systems into cyber-physical systems (CPS).
Concrete examples are production systems, their sub-systems, and components in
the process and manufacturing industry.
    In addition to the consideration of quality, time, and cost targets, the flexibility
of these systems in operation is increasingly coming to the fore. At the same time,
the variety of product types is increasing and the batch sizes are becoming smaller.
Short-term changes in orders as well as changes in supplies are necessarily to be
taken into account. Instead of supplementing classic automation, Digital Twins are
increasingly being named and realized to fulfill these functions. In this chapter,
we consider Digital Twin applications that have simulation-based and data-driven
function modules and can improve operation in various ways (cf. [1, 2]).
    The term Digital Twin has grown historically, introduced in the context of PLM as
to collect the information about a product across its lifecycle [3] and first mentioned as
Digital Twin in a NASA Roadmap in 2010 [4]. Nowadays, it contains all information
about an asset for the entire life cycle and becomes even a synonym for the asset
administration shell in the Industrie 4.0 context [5]. Unfortunately, no consistent
definition is used in practice. It is mostly shaped by the perspective and use of the
respective user.
    In the following, we discuss different lifecycle aspects of Digital Twins and moti-
vate the concept of bundling best practices for simulation-based Digital Twins in the
context of engineering and automation. Then we discuss several novel applications
of the Digital Twin and give concrete examples of their realization. We conclude
with a discussion on the future development of Digital Twins.
In recent years, we have seen an increasing digitization and the triumph of Internet
technologies. Most of the products and systems are created in a digitally supported
development process. Standardized interfaces and data formats are the foundation
of integrated toolchains for the creation and execution of models during all life-
cycle phases [6]. In operation phases, Internet technologies for data acquisition, data
transformation, and data analysis (e.g., by artificial intelligence (AI) methods) have
become cheap, easy to install and use, and have been commoditized. The pervasion
of virtual and augmented reality (VR/AR) into daily-life applications, called Meta-
verse, is currently creating a demand for interactive and intrusive solutions. All this
leads to an increased interest and awareness of Digital Twins and their purpose and
goal.
On the Importance of Simulation and Digital Twin for Industrial …                         97
functional parameters of the product with the capabilities of the production system.
On the other hand, the Digital Twin of the production line can be used to test out
different operating/manufacturing strategies. There is also a certain analogy in the
process industry. Here, too, the behavior of the process plant can be tested and
validated.
    The production, however, is a stage in the lifecycle of the product where the
characteristics of the Digital Twin change significantly. Before the production of the
physical asset, only the “theoretical”/desired behavior of the product in a general-
ized environment is considered assuming estimated operating conditions. As soon
as the product is realized, individual aspects of each of the products are added to
its Digital Twin. This could be the manufacturing history like “manufactured on
machine x with quality y”, “used parts/substances from supplier z” or “batch id
number and environment conditions like temperature or humidity”. Especially in
the case of mass products, each of the products carries along its own production
and operational history, even if the basic product description is initially the same.
In [9], this transformation is called instantiation of the Digital Twin: The Digital
Twin prototype (DTP) that was developed during the design becomes a Digital Twin
instance (DTI) of each individual product. Consequently, the content of the initial
DTP is also available as a common set of information in all instances. This common
set of knowledge is complemented by individual aspects of each instance, like the
individual and specific environment conditions or operational modes. Using these
individual data, individualized predictions for each asset can be made. This requires
on the other side that the data for each asset must be treated individually, which can
be done locally—on the edge—or in a more centralized way or in a mixture of both
extremes, depending on the availability of computing resources.
    Based on this set of information/data, executable models can be created to describe
a specific behavior aspect of the asset. Usually, these models are either numerical
simulation models based on a mathematical description of the relevant physics, or
data-based models that are trained with available sensor data. The desired accuracy
of these models widely depends on the necessities of the specific use case. As these
requirements are often opposite to each other, one has to compromise, e.g., execution
speed versus accuracy, resulting in a “good enough” simulation model for the specific
purpose. Therefore, it could also happen that models of the same physical effect are
part of the Digital Twin in different model fidelity.
    Further challenges arise as the model must cope with real sensor data. Measure-
ments include uncertainties that should be considered in the simulation model as
well. And more importantly, the simulation must process the sensor information in
(near) real time to give fast enough results.
    The Digital Twin will lead to novel software applications in operation and service
phases and, for example, through the establishment of edge and cloud architectures
to new business models. During operation, more and more data, not only from local
sensor measurements but also operating data from connected IT systems, are avail-
able and can be considered in these applications. Technologically, this means that
simulation models are also combined with data-based methods and AI techniques.
On the Importance of Simulation and Digital Twin for Industrial …                      99
From a research point of view and to derive technical ideas for the preliminary devel-
opment, we propose a novel simulation-based Digital Twin concept which extends
traditional simulation approaches and improves most likely the usage of Digital
Twins. This vision of the future has developed further in recent years (see [5, 10,
11]) and manifests itself in the following characterization:
   The (in previous papers called next generation) Digital Twin refers to a description
of a component, product, system, infrastructure, or process by a set of well-aligned,
descriptive, and executable models. It is a semantically linked collection of all rele-
vant digital artifacts, including design and engineering data, operational data, and
behavioral descriptions.
100                                                                     S. Boschert et al.
The development and use of technical systems almost always take place in a network
of development partners, suppliers of parts, components, and products, and of
customers who use these systems or in turn integrate them into their systems and sell
them to their customers. At the end of this chain is the end customer who uses or
operates this system. An example illustrates these connections of partners, e.g., the
component could be an electric drive designed for a control valve. This product is
manufactured in the manufacturing plant and installed in a production plant. Or the
valve is used by system manufacturer to manufacture a complete process module. It is
an open, dynamically changing network. After all, information is already exchanged
between partners or suppliers and customers today (and in the past). The Digital Twin
approach will significantly increase the effectiveness and efficiency of information
exchange, as more complex knowledge can be exchanged using models. The artifacts
exchanged can be Digital Twins or, in terms of the description given above (”a set
of well-aligned, descriptive, and executable models”), a subset of digital artifacts.
Regarding the simulation part, the previously mentioned modularization is essential
here, as it improves reuse and thus increases cost-effectiveness. To put it briefly, the
previous fragmented exchange will be replaced by comprehensive puzzle pieces in
the future. In implementation, their unambiguous description of their content and
their interfaces is important to be able to use them as independent simulation-based
Digital Twins or to connect them with other twins (puzzle pieces). In this way, these
Digital Twins are transformed into independent products that are subject to specific
On the Importance of Simulation and Digital Twin for Industrial …                  101
Autonomous systems have become popular for several years, especially through
assistance systems in vehicles. In addition to the application in the automotive
industry, this development direction is also relevant for other complex technical
systems, e.g., in manufacturing and process industries.
   According to Bekey’s definition [16] “Autonomy refers to systems capable of oper-
ating in the real-world environment without any form of external control for extended
periods of time”, autonomous systems are intelligent machines (or systems again) that
execute high-level tasks without detailed programming and without human control.
They know their capabilities and skills, and their internal state, and they use sensors
to perceive their environment and the current situation. They can make decisions
based on scenario analysis on how to proceed. In order to make this happen, the
On the Importance of Simulation and Digital Twin for Industrial …                   103
autonomous systems will need access to models that are as fine-grained as the use
case requires and their own behavior in interaction with their environment in the
real world—these are typical tasks of a Digital Twin [2]. This once again proves the
relevance of the topic.
   Full autonomous behavior of systems will remain a vision for years, but we will
continuously approach this goal. The concept of Digital Twins will be very useful
for this. An intermediate step is the realization of cyber-physical systems (CPS)
which communicate with their environment and can fulfill more complex tasks. The
economic realization of such CPS will require new development methods, and the
Digital Twin approach will support it.
   A typical approach for system development is to start from requirements and define
functional blocks to fulfill them. These functions are further sub-divided into logical
systems which are finally realized in their physical representation. Mechatronic
systems become more and more cyber-physical, which means that the communication
aspect either within the system but also between different systems becomes more and
more important. Therefore, the system under consideration becomes larger and finally
harder and harder to understand in all details. The quest for the optimal system design
becomes challenging. One approach could be to split the full system into several sub-
systems, e.g., functional blocks and optimize their operation. Especially in classical
mechatronic systems like automotive industry this approach delivers good results,
e.g., by sub-dividing the system “car” into many different components which could
be finally produced by a large ecosystem of suppliers, each one specializing on some
components and optimizing the components by themselves. Only by integration of
the components to the full system are the interactions (physical or communication)
between the separate components realized in a pre-defined way—mostly determined
by the initial functional and logical sub-division of the full system.
   However, as the communication aspect becomes more dominant in cyber-physical
systems, it also becomes more challenging to identify components that are mutually
related and could benefit from each other. This becomes obvious at the example of
a logistics system, where on the one side of a conveyor belt the belt is loaded from
different sources, and on the other side, the material must be sorted into different
buckets. If on the one side the sources are running at maximum speed (are optimized
for maximum performance), the sink side could run into trouble as several buckets
get overloaded—even if the sink side runs on maximum as well as some buckets
are full and must be emptied. In this situation, it would be helpful to slow down the
performance of the source side—or modify the mixture in such a way that the sinks
are more evenly accessed—to maximize the overall performance of the system. Here
source and sink form a symbiotic pair by coordinating their behavior to the benefit
of both components and finally the overall system.
   A central part for the realization of such symbiotic effects is to establish a commu-
nication link between the components. This, however, requires that the relevant
components are already identified. Approaches to identify such symbiotic pairs can
be found, e.g., in [17]. The identification of the symbiotic pairs should have an impact
on the design of the control system as well, as a direct communication is established
more easily if the components are controlled by the same controller. Obviously, an
104                                                                     S. Boschert et al.
Only recently has the Industrial Metaverse (IMV) as evolution of the Digital Twin
era emerged. Whereas the term Metaverse is already used in the gaming market, it is
developing in three additional fields. In the consumer market, it represents an avatar-
centric virtual world for socializing and entertainment, whereas in the enterprise
market, it is used in connection with real-time collaboration to facilitate office tasks
employing large language models in the future. In contrast to that, the IMV as the
third field aims at solving real-world problems, increasing productivity, shortening
time-to-market and, thus, adding value to industry. Prognosis of market segmentation
even sees the IMV at around 100 billion US$, outcompeting the other two fields by
a factor of more than 2–3 [20].
   Concretely, the IMV is a space to experience Digital Twins of industrial assets with
the key features immersion, collaboration, and interaction where the combination
goes beyond Digital Twins. The visualization of the Digital Twin in its context and
realistic environment leads to new insights. Meetings between different stakeholders
to jointly review Digital Twins facilitate problem solving where changes can be
directly applied. Moreover, interactive simulation of the Digital Twin behavior and
On the Importance of Simulation and Digital Twin for Industrial …                      105
Fig. 3 Digital Twin of a production line based on a simulation model and its photorealistic
visualization [21]
106                                                                    S. Boschert et al.
compared to conventional and comparable factories [20]. This, in turn, has impact
on space efficiency, required material, and energy consumption. Consequently, the
IMV can be also seen as a key enabler for sustainability.
   Another example will be explained in the example section.
physical and chemical processes in detail in a single simulation model will lead to
high complexity and computational costs. Therefore, it is reasonable to use simpli-
fication methods like homogenized models. To ensure that the selected models are
applicable and correct for describing a specific process, it is necessary to validate
these models with physical tests or high-resolution multi-physics simulations (if
high-fidelity simulations are available for the respective process). For the model
verification, a physical test or high-resolution simulations can be performed for a
representative volume or simplified geometry. The validated models are then used to
simulate and optimize the complex structures and geometries of functional compo-
nents. As such, already during the early design phase a set of specific simulation
models is created that form the initial Digital Twin of the reactor.
    Let us look at the workflow from detailed representations to homogenized models
for 3D simulations of methanol synthesis reactor. The methanol synthesis is a
complex chemical process that can be described by means of kinetic models [23]
received from fitting the measurement results. In the high-resolution multi-physical
simulations, the geometry of the catalyst is reproduced in detail. Therefore, the
properties such as pressure drop or effective thermal conductivity of the catalyst
bed for a homogeneous model are calculated. Similarly, the basic properties of the
porous metallic structures in the reactor are defined. Using the high-resolution simu-
lations and tests the porous materials were analyzed, taking into account the geometry
including all pores and collected information. Then this information is used to find
out the parameters for the homogenized models like the Darcy-Forchheimer model
for pressure drop [24, 25]. The difficulty so far lies in the calculation of the dynamic
capillary pressure with detailed pore geometry, since this simulation is transient,
requires a high computational effort, and depends on the contact angle, which does
108                                                                          S. Boschert et al.
not always remain constant during the solid–liquid interaction. Currently, the high-
resolution multi-physics simulations are not available for the porous structures in
our chemical reactor. As a first approximation, the Leverett model [26] was used
to describe the capillary behavior in a porous structure. For the description of the
turbulent flow in the reactor, Reynolds-averaged Navier–Stokes equations (RANS
equations) with k-ε turbulence model [27, 28] are used, which provide results of
sufficient accuracy for given flow boundary conditions in the reactor. To describe the
phase transition a model with Eulerian averaging of the transport equations for the
additional phase interaction was used [29, 30]. This model solves the conservation
equations and assumes a common pressure field for the phases. It is important to
mention that the main goal of the models is to describe the process quantitatively
and not the micro-scale process reproduction (Fig. 5).
    The simulation models that use homogenization of small-scale phenomena and
parametrization of geometry provide results during reasonable time for the industry
and allow to build an automatic circuit “CAD-simulation-optimization-CAD” to get
an optimal novel design of functional structures.
    The created 3D multi-physics simulations for the understanding of detailed
processes and geometry optimization can be also applied to determine the missing
input data and boundary conditions like pressure drop in the catalyst bed and heat
transfer coefficients for one-dimensional simulations (1D simulations). The 1D simu-
lations allow not only to calculate the chemical transformation in a reactor under given
conditions but also can consider the reactor as a part of a chemical plant and define
the operation points for an entire facility. The local 3D simulation of the reactor
can be included in an overall system model to optimize the operation of the facility.
On the other hand, the system model provides more accurate boundary conditions
Fig. 5 Example of the connection between high-resolution simulations and modeling for methanol
synthesis reactors
On the Importance of Simulation and Digital Twin for Industrial …                   109
for the local model. Therefore, the example of the methanol synthesis shows that
the development of design for a functional component is an iterative process using
Digital Twin models of varying degrees of scale. These models improve during this
iterative process and are then the base for a proper operation of the whole plant.
For a health monitoring application, a specific aspect of Digital Twin is utilized that
focuses primarily on modeling and simulating the behavior of a physical system or
process (Behavior Twin). This can be used to predict how the system will behave in
different scenarios and to support decisions or optimizations. This approach is applied
at different stages of the development to ensure the functionality and reliability
of complex systems, particularly in operation and service for health monitoring of
critical components.
    The knowledge of the health status of electrical drive train components, such as
converter and motor, is important for safe operation and for planning the service
activities. To calculate the aging, knowledge of the temperature of critical compo-
nents (converters electronic, bearing, stator insulation) during the operation is essen-
tial. Using numerical simulation models, the calculation of health and remaining
useful life by processing of operational data and estimation of temperatures of crit-
ical components becomes possible. It eliminates the need for direct measurements,
which are often unfeasible due to limited access to the components and the huge
volume of operational and sensor data generated.
    The workflow for monitoring the health of the drive train is shown in Fig. 6. This
process consists of several steps which contribute to the development and deployment
of a comprehensive health monitoring service. The first step involves the creation of
detailed models for each critical component of the drive train: (a) the application, (b)
the motor, and (c) the converter. These models describe components’ behavior and
consist of various digital artifacts, including system architectures, CAD representa-
tions of sub-components, materials, and control models. In the next phase (system
integration), the component models are combined to a model of the entire drive train
system. This step allows to simulate the interactions between the different compo-
nents, evaluate the system’s performance, and calculate the temperature of critical
components. To create a robust health monitoring service, the aging models are
integrated into the system model. These models consider the wear that drive train
components undergo over time and allow to estimate the components’ health and
remaining useful life.
    It is important to know that these models are developed for a specific application
and need to be precisely evaluated to ensure that they are applicable for another
scenario. If a model does not meet the defined requirements, it is necessary to create
a new model or modify an existing model.
    The execution of the developed service usually takes place on edge computers
(d) close to the data source. The most important advantages are low transmission
110                                                                                     S. Boschert et al.
                                                                                  Service development
  a)                   b)                   c)                d)
                                                                                  e)    Aging model
                                                              Model integration
                Development of component models
latency, reliability, and security. The other option is to run it as a cloud service. The
main reasons for this are scalability, flexibility in updates and maintenance.
   In summary, the Digital Twin technology provides an innovative and physics-
based solution for the calculation of stress, temperature, and damage levels in compo-
nents without direct measurements. By applying the correlation between component
use, characteristics, and operating conditions, these predictive maintenance models
provide the way for safer and more efficient operations as well as the development
of more robust components.
the project is to create a collaborative and open data ecosystem for the automotive
industry. Digital Twins are a core element to realize the provision and exchange of
data [34].
    The information required for the exchange between partners in the data ecosystem
is described as semantic data models. For transmission, they are filled with the specific
data to be exchanged, i.e., instantiated. This principle is implemented via asset admin-
istration shells [35]. In Industrial Digital Twin Association (IDTA) terminology, these
data models correspond to sub-models. These sub-models are supplemented by infor-
mation about the object to which this data relates. Both parts together form the asset
administration shell. In the understanding of IDTA, the asset administration shell and
Digital Twin are synonyms [6]. Digital Twins are therefore the means of transport
for software solutions in data ecosystems. They are used, for example, to exchange
information along the supply chain, as the next example illustrates.
Online Control and Simulation (OSim) is the name of a use case in the German public-
funded project Catena-X. The goal of the research development is a software solution
which allows a distributed, collaborative material flow simulation of production and
logistics. The strategic goal is the increased resilience of supply chains by faster,
more precise reactions in case of short-term disruptions [36]. OSim is one of five use
cases in the Catena-X Business Domain Resiliency [37].
112                                                                       S. Boschert et al.
   Supply chains are subject to various disruptive factors, and it is important to react
flexibly and make effective adjustments. An obstacle to this is the often insufficient
exchange of information between the individual partners in the chain. Even if indi-
vidual logistics or production partners carry out their planning locally optimally,
without overarching coordination this can lead to inefficient supply chains.
   The OSim solution approach aims to simulate the behavior of the entire supply
chain, including the effects of short-term disruptions or changes, at the material flow
modeling level [38]. The simulation is not carried out in a holistic, monolithic model,
but as a distributed simulation in which each partner simulates their processes and
sequences. The exchange of information with suppliers and customers necessary for
the respective simulation takes place via standardized data models. It maintains the
data sovereignty of the partners and guarantees compliance with antitrust regulations.
   Through collaborative simulation, it becomes possible to react to changes in
advance and improve efficiency across the entire supply chain by the increased
observation horizon.
   All partners involved in the supply chain simulate their respective processes and
actions (production or logistical sequences) based on current information from the
shop floor and its IT systems (e.g., manufacturing execution systems). The results of
the simulation, essentially delivery data, and quantities of parts produced or trans-
ported, are passed on to the immediate successor in the supply chain, who in turn
carries out a simulation of their material flow processes based on this possibly
changed information. In OSim, a “horizontal” exchange of information across the
supply chain is combined with a “vertical” exchange of the material flow simulation
of its production or its logistics processes (e.g., transport times). Through iterative
simulations by all partners and information exchange upstream and downstream,
improved efficiency of the entire supply chain is achieved in a collaborative way.
The core of this approach is an OSim manager, which is operated by every partner
involved. The OSim manager is a software application that records all the information
that is necessary for the local execution of a material flow simulation and supplies it to
the locally executed simulation. After the simulation has been carried out, the OSim
manager provides the results of the simulation to the partners again. For the secure
and data-sovereign transmission of information, the Eclipse Data Space Connector
(EDC) is used in accordance with the Catena-X regulations [39].
   The executed OSim manager with its connected local simulations forms a Digital
Twin of the material flow of the entire supply chain. Also, every exchange of infor-
mation between the partners uses (in the IDTA understanding) a Digital Twin, which
essentially contains the results of the simulation.
Use cases for the Industrial Metaverse are numerous and manifold. Although it
still represents a vision, Siemens and Freyr have already showcased in [40] how
On the Importance of Simulation and Digital Twin for Industrial …                     113
Fig. 8 Digital Twin of a battery plant including production and assembly lines [40]
manufacturers can work in the future based on the example of a battery production
site, see Fig. 8.
    By creating the Digital Twin including an Industrial Metaverse experience of
a factory that hasn’t been built yet, confidence can be built among stakeholders
including investors. Being immersed into the future production site, seeing the
production lines virtually commissioned, and having transparency on the perfor-
mance of the whole plant creates trust in the capability of the corresponding company
and reduces the risk of wrong investments significantly.
    Moreover, the IMV provides the means to not only display KPIs from the whole
plant down to the component level but also to optimize them. Here, simulation plays a
crucial role. For example, in the critical mixing process (see Fig. 9), the mixing degree
can be simulated by means of CFD (computational fluid dynamics) and particle simu-
lation to finally optimize the process time impacting the overall performance of the
plant. If this is not enough, then even the machine itself or machine components can
be optimized. Simulation and design optimization enable dedicated improvements
like the geometry of the stirrer. Thus, the IMV allows for virtually testing different
parameters before the plant is built or without compromising the actual production
of an existing plant. This becomes especially important in the operational phase of
a plant when anomalies are detected, and a root-cause analysis can be performed in
the IMV. Here, different stakeholders from different locations can virtually meet in
the plant, analyze operational data, and test corrective actions before applying them
in the real plant. With this approach, the costly downtimes of a plant can be reduced
to a minimum.
    There are many more use cases where the IMV creates an added value in a factory.
In the assembly (see Fig. 9 on right) health and safety issues can be evaluated such as
the interplay between robots and human workers or test the ergonomics by simulation.
As energy consumption and the corresponding CO2 equivalent became a major KPI
in production, the IMV of a plant can also be used to create transparency and to
optimize scheduling to additionally save costs by making use of changing energy
prices.
114                                                                          S. Boschert et al.
Fig. 9 Simulated mixing process allows for optimization of mixing time and mixing degree [20]
The term Digital Twin is now being used for more than ten years. Its roots lie in
simulation, other influences came and continue to come from areas such as data
analytics, the Internet of Things (IoT), and Industry 4.0. The latter claims to equate
the terms Digital Twin and asset administration shell [6].
    A comparison between the definition given here by the authors and the Asset
Administration Shell shows great similarities. Information and data from different
sources are used and integrated. Knowledge capturing takes place; data and
simulation-based models play a central role here. Both approaches address the entire
lifecycle of assets, from design and development to operation and maintenance.
Furthermore, there is a clear relationship to the counterpart of the Digital Twin—
the real twin. These similarities justify using both terms synonymously and further
convergence will increase as further sub-models of the AAS are defined. To evaluate
the future development of the Digital Twin approach, it makes sense to take a focused
look at the core properties of both terms and the approaches behind them. The Asset
Administration Shell is a standardized framework that provides a digital represen-
tation of a physical asset. It serves as a comprehensive information model, which
includes in particular administrative and business-related information on the physical
asset. The framework defines a common language designed for seamless integra-
tion between different assets and interoperable communication between systems and
devices. The Digital Twin approach emphasizes the creation process of the digital
representation. This representation includes relatively detailed and dynamic behav-
ioral models consisting of detailed analysis, optimization, and automated support of
users in their different tasks in all life cycle phases. A Digital Twin refers not only
On the Importance of Simulation and Digital Twin for Industrial …                    115
to physical assets, but also to planned real objects (from small components to large
infrastructures), technical and business processes and people.
    The term Digital Twin is associated with two aspects that must be viewed in a
differentiated manner. On the one hand, this is the representation of the real coun-
terpart itself and, on the other hand, the use and application of the Digital Twin. The
first aspect includes elements of a scientific discipline that pursues technical improve-
ment of the models used and works on further developing the methods required for
this. The second aspect addresses the creation of software applications that directly
provide the Digital Twin as a concrete solution for a task. The types of solutions
range from simple tools to fully automated IT solutions.
    This leads to a future vision of the Digital Twin that is characterized by several
elements. The importance of the models used will continue to increase. These must
be easier to combine in their usage and therefore more modular in their methodolog-
ical basis. The models must meet typical product characteristics. These are explicit
descriptions of the model contents and model limitations as well as assurances about
model quality in terms of maturity and stability. In order to ensure economic use,
these models must be easier to develop and maintain. This refers on the one hand to
the further use of the models or the Digital Twin for variants of the planned or real
counterpart and on the other hand to the successor generation of the real twin. This in
turn requires better integration into development tools and existing or new software
solutions for operation and service, which is also a task in the Digital Twin tech-
nology area. These more scientific and technical future aspects of the Digital Twin
continue approaches that were started in the 1980s with the CAx methods and are
now finding a future with the Industrial Metaverse. On the application perspective of
Digital Twins, another important element is the efficient generation of mostly soft-
ware applications which allow the direct usage of a Digital Twin for concrete tasks.
The Asset Administration Shell framework plays a crucial role in implementing
these goals. It enables the management of Digital Twins and their integration into
various applications. However, other objectives require improvements on technical
and scientific elements. On the Digital Twin application perspective, it is expected that
specific application solutions besides the AAS will be continuously developed which
are independent and directly marketable products. The Digital Twin and its further
development offer numerous opportunities and potential for companies in various
industries. By using models that have modularity and clear product descriptions as
well as integrating appropriate tools and applications, new uses can be opened up
and efficiency increased. The Digital Twin is therefore an important technology for
the future of industry to realize better insights and decision making for the planned
or existing real counterpart.
116                                                                               S. Boschert et al.
References
 1. Kuehn W (2018) Digital twins for decision making in complex production and logistic enter-
    prises. Int J Des Nat Ecodynamics 13(3):260–271. https://doi.org/10.2495/DNE-V13-N3-
    260-271
 2. Rosen R, von Wichert G, Lo G, Bettenhausen KD (2015) About the importance of autonomy
    and digital twins for the future of manufacturing. IFAC-Pap 48(3):567–572
 3. Grieves M (2002) Completing the cycle: using PLM information in the sales and service
    functions [slides]. SME Management Forum, Troy, MI
 4. Piascik R, Vickers J, Lowry D et al (2010) Technology area 12: materials, structures, mechanical
    systems, and manufacturing road map. NASA Office of Chief Technologist
 5. Wagner C, Grothoff J, Epple U, Drath R, Malakuti S, Grüner S, Hoffmeister M, Zimmermann P
    (2017) The role of the Industry 4.0 Asset Administration Shell and the Digital Twin during the
    life cycle of a plant. In: Proceedings of the 2017 IEEE conference on Emerging Technologies
    Factory Automation (ETFA 2017), Limassol, Cyprus
 6. Rosen R et al (2021) Die Rolle der Simulation im Kontext des Digitalen Zwillings: Virtuelle
    Inbetriebnahme-Modelle als Mittler zwischen Phasen und Anwendungen? atp magazin, Bd 63,
    Nr 04, S. 82–89
 7. Boschert S, Rosen R (2016) Digital Twin—the simulation aspect. In: Hehenberger P, Bradley
    D (eds) Mechatronic futures. Springer International Publishing, Cham, pp 59–74
 8. Drath R, Malakuti S, Grüner S, Grothoff J, Wagner C, Epple U, Hoffmeister M, Zimmer-
    mann P (2017) Die Rolle der Industrie 4.0 „Verwaltungsschale“ und des „digitalen Zwillings“
    im Lebenszyklus einer Anlage - Navigationshilfe, Begriffsbestimmung und Abgrenzung. In:
    Tagungsband zur Automation 2017. Langfassung auf Tagungs-CD (12 Seiten), VDI-Verlag,
    Baden-Baden. ISBN 978-3-18-092293-5
 9. Grieves M, Vickers J (2017) Digital Twin: mitigating unpredictable, undesirable emergent
    behavior in complex systems. In: Kahlen FJ et al (eds) Transdisciplinary perspectives on
    complex systems. Springer, Heidelberg
10. Boschert S, Rosen R, Heinrich C (2018) Next generation digital twin. In: Horvath I et al (eds)
    Proceedings of TMCE 2018, S 209–218
11. Rosen R, Fischer J, Boschert S (2019) Next generation digital twin: an ecosystem for
    mechatronic systems? IFAC-Pap 52(15):265–270. https://doi.org/10.1016/j.ifacol.2019.11.685
12. Barth M, Kübler K, Heinzerling T, Rosen R, Jäkel J (2020) Eine systematische Bewertung der
    Qualität von Simulationsmodellen: Identifikation und Clustering von Qualitätskriterien. atp
    magazin (06–07), S 68–75
13. Barth M et al (2021) Eine systematische Bewertung der Qualität von Simulationsmodellen,
    Teil 2: Bewertungsmetrik und Validierung, atp magazin, Bd 63, Nr 04, S 56–63
14. Kalhoff J, Borchers E, Rosen R (2021) „Diskussionspapier: Production System Life-
    cycle Management im Wandel der Fahrzeugindustrie“, Plattform Industrie 4.0, avail-
    able at https://www.plattform-i40.de/PI40/Redaktion/DE/Downloads/Publikation/Diskussio
    nspapier_kopa35c.html. Last access 04 June 2021
15. Bröcker S et al (2021) Prozesssimulation – Fit für die Zukunft?, Positionspapier der
    DECHEMA und VDI. Available at: https://dechema.de/Positionspapier_Prozesssimulation-
    path-123212,124930.html. Last access 04 June 2021
16. Bekey GA (2005) Autonomous robots: from biological inspiration to implementation and
    control. The MIT Press. ISBN 9780262025782
17. Zeman K et al (2020) Symbiotic mechatronics: an alternative view on complex systems. In:
    Proceedings of TMCE 2020, 11–15 May 2020, Dublin
18. SG AAS Teilmodell Simulation, Plattform Industrie 4.0 (2021) „Interoperabilität hoch zwei
    für Simulationsmodelle. Available at https://www.plattform-i40.de/PI40/Redaktion/DE/Dow
    nloads/Publikation/Interoperabilit%C3%A4t-hoch-zwei.html. Last access 04 June 2021
19. Fischer J et al (2021) Assistenzsysteme der nächsten Generation für Produktionssysteme, VDI
    conference Automation 2021, Paper accepted
On the Importance of Simulation and Digital Twin for Industrial …                              117
20. MIT Technology Review Insights (2023) The emergent industrial metaverse. https://ass
    ets.new.siemens.com/siemens/assets/api/uuid:260d1ead-250b-45fb-b32f-caa9924c1ddd/The-
    Emergent-Industrial-Metaverse.pdf
21. https://www.siemens.com/global/en/company/digital-transformation/industrial-metaverse/
    what-is-the-industrial-metaverse-and-why-should-i-care.html
22. Kalz KF, Kraehnert R, Dvoyashkin M, Dittmeyer R, Gläser R, Krewer U, Reuter K, Grun-
    waldt J-D (2017) Future challenges in heterogeneous catalysis: understanding catalysts under
    dynamic reaction conditions. ChemCatChem 9:17–29
23. Lacerda de Oliveira Campos B, Herrera Delgado K, Pitter S, Sauer J (2021) Development of
    consistent kinetic models derived from a microkinetic model of the methanol synthesis. Ind
    Eng Chem Res 15074
24. Darcy H (1856) Dalmont V (ed) Les Fontaines Publiques de la Ville de Dijon: Exposition
    et Application des Principes à Suivre et des Formules à Employer dans les Questions de
    Distribution d’Eau (Paris)
25. Forchheimer P (1901) Wasserbewegung durch Boden. Z Ver Dtsch Ing 45:1781–1788
26. Leverett MC (1941) Capillary behavior in porous solids. Trans AIME 142:151
27. Lien FS, Chen WL, Leschziner MA (1996) Low-Reynolds number eddy-viscosity modelling
    based on non-linear stress-strain/vorticity relations. In: Proceedings 3rd symposium on
    engineering turbulence modelling and measurements, 27–29 May, Crete, Greece
28. Teli SM, Mathpati CS (2020) Computational fluid dynamics of rectangular external loop airlift
    reactor. Int J Chem Reactor Eng 18(5–6):20200009
29. Drew DA, Passman S (1998) Theory of multicomponent fluids. Springer, New York
30. Ishii M (1975) Thermo-fluid dynamic, theory of two-phase, eyrolles
31. Plattform Industrie 4.0. Manufacturing-X. Available at https://www.plattform-i40.de/IP/Nav
    igation/EN/Manufacturing-X/Manufacturing-X.html. Last access 19 Oct 2023
32. Plattform Industrie 4.0. White Paper on Manufacturing-X. Available at https://www.plattform-
    i40.de/IP/Redaktion/EN/Downloads/Publikation/Manufacturing-X_long.html. Last access 19
    Oct 2023
33. Federal Ministry for Economic Affairs and Climate Action (BMWK). Projektsteckbriefe. Avail-
    able at https://www.bmwk.de/Redaktion/DE/Downloads/P-R/projektsteckbriefe.pdf?__blob=
    publicationFile&v=1. Last access 19 Oct 2023
34. Catena-X (203) Semantic layer/digital twins. https://catena-x.net/en/offers-standards/sem
    antic-layer. Last access 19 Oct 2023
35. IDTA, Industrial Digital Twin Association, information available at https://industrialdigitalt
    win.org/en/technology. Last access 19 Oct 2023
36. Eclipse Tractus-X. Osim Kit. Available at https://eclipse-tractusx.github.io/docs-kits/kits/
    OSim%20Kit/Adoption%20View%20OSim%20Kit. Last access 19 Oct 2023
37. Eclipse Tractus-X. Business domain resiliency. Available at https://eclipse-tractusx.github.io/
    docs-kits/next/Resiliency/. Last access 19 Oct 2023
38. Wolfrum P, Haager D, Isinger T, Rosen R, Kutzler T (2023) Kollaborative Materialflusssim-
    ulationen zur Steuerung der Verknüpfung von Supply Chain und Shopfloor - ein Catena-X
    Use-Case. In: Automation, Baden-Baden
39. Catena-X, EDC—The Central Component. Available at https://catena-x.net/en/offers/edc-the-
    central-component-1. Last access 19 Oct 2023
40. Siemens FREYR Industrial Metaverse showcase. YouTube. https://www.youtube.com/watch?
    v=LQyKc6SI5u8
A Formal Framework for Digital Twin
Modeling, Verification, and Validation
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2024          119
M. Grieves and E. Hua (eds.), Digital Twins, Simulation, and the Metaverse, Simulation
Foundations, Methods and Applications, https://doi.org/10.1007/978-3-031-69107-2_6
120                                                                     M. K. Traore et al.
    The Digital Twin (DT) concept has risen as a pivotal strategy, landing in top
strategic technology trends. It revolves around the concept of mobilizing a model
in place of a system, ensuring continuous synchronization between the model and
the actual system. This synchronization serves to reflect real-world events on the
model, enabling the assessment of management initiatives on this ever-updated arti-
fact before implementation. Hence, the model is more than a simple representation,
evolving into a digital counterpart intricately linked to the specific system in question.
NASA pioneered the system-pairing approach, simulating situations in its space-
craft from distant command centers to guide astronauts in interventions. Notably, the
famous Apollo 13 mission in 1970 relied on a system-pairing technique but didn’t
involve a DT; instead, it utilized two Physical Twins (one in space and the other on
land).
    The term “Digital Twin” was introduced by [1], acknowledging its underlying
principle foreseen by [2]. In the context of product lifecycle management, the con-
cept evolved through various stages: Mirrored Spaces Model [3], Information Mir-
roring Model [4], and finally as Digital Twin [5]. The concept is defined as “a set
of virtual information constructs that fully describe a potential or actual physical
manufactured product, from the micro-atomic to the macro-geometric level” [6].
This data-centric view shifted to behavioral aspects in [7], defining a DT as “an
integrated multi-physics, multiscale, probabilistic simulation of an as-built vehicle
or system that uses the best available physical models, sensor updates, fleet history,
etc., to mirror the life of its corresponding flying twin”. From the simulation per-
spective, this approach is disruptive, as simulation experiments are based on current
information from the system rather than assumptions. Simulation experiments rely
on the current information available from the system rather than assumptions [8, 9].
In this application, DT not only serves a representational function but also proves
effective in predicting the expected behavior of the system [10, 11]. Consequently, a
DT is not merely a comprehensive model of the system it represents but a collection
of integrated sub-models reflecting different aspects of the system [12]. Additional
dimensions have emerged, including the use of DT for prognostic and diagnostic
activities [13, 14], as well as real-time optimization [15, 16].
    Nowadays, DT applications are diverse and extend across various industries,
including automotive [17], avionics [18], aerospace [7], energy [19], manufacturing
[8], health care [20], and services [21]. Within industrial contexts, DTs play a crucial
role in predictive maintenance of equipment, enhancing asset reliability and safety,
and optimizing both product design and process operations. In health care, the DT
approach holds the potential to revolutionize medical treatments by tailoring them
to the individual patient’s needs rather than relying on generalized approaches. Fur-
thermore, DTs contribute to the servitization trend by enabling companies to monitor
their products during customer usage, enhancing the overall customer experience.
    DTs can be divided into three [22] main types: Digital Twin Prototype (DTP)
oriented for Product Lifecycle Management, Digital Twin Instance (DTI) oriented
for individual product that replicates one single physical or numerical ToI, and Digital
Twin Aggregate (DTA) that is an aggregate of several DTIs. In this chapter, we will
A Formal Framework for Digital Twin Modeling, Verification, and Validation                 121
Fig. 1 High-level view of Digital Twin in interaction with both human and physical sides
mainly focus on DTI type, but contributions can be applied to both DTP, DTI, and
DTA.
    Despite numerous research and development initiatives, the verification and vali-
dation (V&V) of DT remains a significant scientific challenge. To address this obsta-
cle, there is a pressing need for a comprehensive framework. This chapter responds
to this need by introducing a potential framework designed to achieve effective
DT verification and validation. As a second contribution, we propose three dif-
ferent approaches to reach the goal of DT V&V which can be applied to all DT
types: simulation based, software engineering based, and formal method based. This
approach extends to modeling, verifying, and validating the interoperability middle-
ware, specifically the Internet of Things (IoT) connection between the system and
its corresponding DT.
    Thus, we can observe two main interacting elements: a physical or numerical
model of a system of interest, sometimes called Physical Twins, or Replica Twins
[6] in literature; however, we prefer to refer to it as the Twin of Interest (ToI),
enabling us to incorporate both physical models and non-physical ones, such as
numerical models. Perpetual synchronization between the TOI and the Digital Twin
is maintained, based on data tracked in the TOI. Considering that humans are in most
cases present in systems as stakeholders or decision-makers, e.g., health care [20],
smart cities [23], services [21], cyber physical and human systems [24], we can add
a third interacting element, the human.
    Figure 1 provides an overview of the value chain, emphasizing key components
necessary for deploying an effective Digital Twin (highlighted in blue). The triangle
comprises three primary stakeholders: the physical (component to be replicated), the
digital (core Digital Twin), and the human (users and managers). On the physical
side, representing a cyber-physical system, the integration of cyber and physical
components allows operations executed by actuators, with data collected by sensors
and transmitted through a network. The digital side hosts the Digital Twin, receiving
data into models, enabling decisions sent back to the system or used by the user or
manager for further management decisions. The human side accommodates major
122                                                                     M. K. Traore et al.
stakeholders, including users and managers, utilizing the Digital Twin to explore use
cases and develop strategies.
    Digital–physical interactions focus on the symbiotic relationship between the
physical part and the Digital Twin infrastructure, addressing scientific and techno-
logical challenges. These include designing the end-to-end data circuit, conceptualiz-
ing data needs, integrating heterogeneous data sources, and addressing requirements
for IoT liability and efficiency. Additionally, edge computing is employed for faster
computations related to real-time data streams, such as algorithms transforming raw
data for the Digital Twin’s use.
    Digital–human interactions involve applying advanced technologies like Vir-
tual/Augmented Reality and Web/Mobile approaches to bridge the reality gap in
interfacing humans with the Digital Twin. This addresses issues like using Metaverse-
type technologies for the last mile to end-users, enhancing decision-making, and for-
malizing “cognitive interoperability” in the Metaverse through social interactions.
Digital-based human–physical interactions explore how Digital Twin models and
services impact the relationship between the Twin of Interest/physical subject and
the human subject: user (behavioral changes) or decision-maker.
    The blue component of Fig. 1 (the Digital Twin part) can be zoomed in and illus-
trated as Fig. 2 according to an understanding framework highlighting the tree main
nodes-graphs common to all Digital Twins. The intertwining of these three nodes is
commonly referred to as a DT. However, a DT can be assimilated at different scales:
a process perspective (macro) or a product perspective (micro). For instance, in the
case of a production system, the DT at a process level could represent an entire indus-
trial production line. The feedback from the twin would include information about
the flows moving through the factory, with the twin providing guidance for managing
this chain. In the same context, at a different scale, from a “product” perspective,
the twin could be the DT of an engine in the production line. This distinction can
be illustrated with an example from a service system. The twin of a process (macro)
might be the DT of a smart city, while the twin of a product (micro) could be that of a
patient in the healthcare domain. This flexibility in scale allows the concept of Digital
Twins to be applied broadly, capturing both the holistic view of entire systems and
the detailed perspective of individual components or products within those systems.
In this chapter, we present a diagram illustrating the core DT. Figure 2 illustrates its
composition as three nodes: data, models, and services.
   Figure 2 is the understanding framework, the symbiotic association of data, mod-
els, and services. In the Digital Twin, data stored and used are not only real-time
information from sensors but also draw from various legacy sources, aggregating
historical data, user equipment (e.g., smartphones, embedded cameras), open data,
and Internet-based APIs (e.g., Google Maps). Digital Twin models are structured into
modules, each focused to specific objectives and designed to address particular ques-
A Formal Framework for Digital Twin Modeling, Verification, and Validation         123
tions (e.g., a physical system evolution in its environment under specific conditions,
the impact of adding/modifying infrastructures, identification of upcoming security
vulnerabilities, etc.). Modules are developed across specific domains depending on
the subject of interest.
   As we can see on Fig. 3, one of the modules is the services provided by the Digital
Twin, also defined considering the needs and the area of the subject twinned. They
can be separated in four main categories: (1) monitoring data according to geometric
(2D/3D) or non-geometric models, or directly from data [25], (2) diagnosis is based
on regression, statistical, and stochastic models to identify and prevent events [26],
(3) prognosis capability [27] allows forecast or predicts future states, conditions, or
events based on the analysis of current and historical data and simulation models [7,
13], and (4) prediction ability to anticipate or forecast future outcomes, behaviors,
or trends based on simulation and optimization models [16].
   Service node component is based on the model node which contains a collection
of integrated sub-models reflecting different aspects of the system twinned. Those
models can be geometric or non-geometric models for monitoring, statistical or
stochastic models for diagnosis services, simulation models and optimization models
for prognosis and prediction services. The use of these models fed by data enables
service execution. Moreover, these models can have the ability to update through
data–model interaction. Data node component serves as the foundational pillar that
includes all relevant information generated by real-time sensors, legacy data sources,
and data results from other services. The data node is crucial for fueling models;
thus, the quality and diversity of these data determine the reliability of predictions,
diagnostics, and analyses performed by the Digital Twin. By integrating varied data,
ranging from real-time conditions to historical data, the “Data” node provides a robust
foundation for prediction, diagnostics, and optimization services, contributing to an
informed and effective decision-making process.
124                                                                     M. K. Traore et al.
   Moreover, the complexity shown in the figure includes various layers of inter-
actions involved in the Digital Twin’s overall process. The data node of the Digital
Twin emerges as the gateway facilitating seamless communication with the physical
realm. Concurrently, the services’ node acts as the portal enabling the commissioning
(activation and utilization) of the Digital Twin by the urban decision-maker. On the
other hand, citizen engagement finds its avenue through the models’ node, shaping
a dynamic and interactive ecosystem within the smart city’s digital representation.
This intricate network of nodes and interactions highlights the sophisticated web of
relationships and functionalities within the Digital Twin framework.
   Having explored the structure and high-level architecture of the Digital Twin in
this chapter, the next logical step is to dive into the modeling and formal specifi-
cation aspects. Understanding the composition and intricacies of the Digital Twin
sets the stage for a more detailed examination of its modeling and standardization in
the following chapter. This transition is pivotal as it emphasizes the need for a sys-
tematic and formalized approach to enable the subsequent phases of verification and
validation. With a comprehensive understanding of the Digital Twin’s composition,
the focus now shifts toward establishing robust models and standardized practices,
paving the way for a thorough examination of its capabilities and functionalities.
Formal specification brings valuable advantages. Firstly, it clarifies the intended con-
cept, making it less open to different interpretations. Secondly, it allows the creation
of systematic approaches to generate models that can be executed. Thirdly, it pro-
vides a possibility for symbolic manipulation to formally check the consistency of the
specification and its alignment with certain requirements: verification and validation
(V&V). Thus, the need for methods and tools for modeling and specifying Digital
Twins is a crucial point addressed in this document. Without a formal description of
these twins, it becomes impossible to perform formal verification and specification
to ensure their alignment with the case study, regardless of the application domain.
    The scientific community agrees that there is a significant difference between tra-
ditional models and digital twins. In [29–32], authors provided summaries of differ-
ent definitions of Digital Twins and proposed generalized definitions and distinctive
characteristics to differentiate Digital Twins from other models.
    However, it is challenging to find traces of research that focus not on modeling
Digital Twin models, but on modeling the Digital Twin itself. Some attempts have
been made to achieve this goal, but they are often not sufficiently thorough. In [33],
authors identified and categorized different models within Digital Twins into six
categories: application domain, hierarchy, discipline, dimension, universality, and
functionality. However, this contribution mainly focuses on analyzing models within
Digital Twins, without addressing the overall model of Digital Twins itself. The same
author, in [34], proposes to deconstruct and investigate the digital models into six
distinct aspects: model construction, model assembly, model fusion, model verifi-
A Formal Framework for Digital Twin Modeling, Verification, and Validation            127
cation, model modification, model management, but once again, this work focuses
on the models inside the Digital Twin, not the general model of the DT itself. A
similar work has been done in [35] where author in this paper reviews and analyzes
several methods and modeling techniques used in DT domain in order to classify
modeling methods in groups, e.g., physics-informed ML; data-driven modeling; sys-
tem modeling; physics-based modeling; geometric modeling, etc. We will see that
this approach brings an interesting perspective because depending on the application
domain, modeling methods differ, which implies variations in verification and vali-
dation methods. In [22], a DT is presented as a composition of basic components that
provide on one hand, basic functionalities (identification, storage, communication,
security, etc.), and on the other hand, an aggregated DT is defined as a hierarchical
composition of other DTs, which is a first trace in the literature of a model not internal
to the DT, but of the DT itself. The article proposes a reference architecture of DT
using Automation ML through model-driven engineering approach.
    Based on the state of the art, we can observe that DT modeling domain covers
two main aspects. First aspect is “modeling for DT", which means the design of the
model within the Digital Twin, or the architecture of the Digital Twin. The second
aspect is “modeling of the DT itself”, where there are very few articles, if any, that
talk about modeling the Digital Twin itself.
    Formal specification methods offer significant advantages for Digital Twin mod-
eling [36], regardless of the application context. However, as we can see, there is
no standardized approach for the formal modeling of Digital Twins. Consequently,
facilitating the verification and validation (V&V) of Digital Twins is not a straight-
forward task. This gap in standardization shows the need for a systematic and formal
framework that can be universally applied to ensure the accuracy and reliability of
Digital Twins across diverse domains. Establishing such a standard would contribute
to advancing the field, providing a common ground for researchers and practitioners
to enhance the V&V processes and overall robustness of Digital Twin implementa-
tions.
    To formally specify a DT, we use a system-theoretic approach introduced in 1976.
In this approach [37], the DT model is considered a black box with inputs and outputs
that connect it to the environment through sensors and actuators. Zeigler’s paradigm
includes fundamental entities in the modeling and simulation field. These entities
are represented by Fig. 5 left side, consisting of the source system (the system under
study), the context, the model, and the experimental frame (EF). Please note that the
simulator is not shown here. The source system provides behavioral data, the context
defines the observation conditions, the model represents the system abstractly using
rules or equations, and the EF, acting as an abstract representation of the context,
couples with the system model to generate data under specific conditions. The sim-
ulator, although not illustrated, is the automaton capable of executing instructions
from the resulting coupled model.
    In contrast to a conventional model (left side in Fig. 5), which is built as an
abstract representation of a source system for experimentation under specific condi-
tions defined by an Experimental Frame (EF), a DT model (right side in Fig. 5) is
continuously refreshed with data from the source system. This ongoing update neces-
128                                                                     M. K. Traore et al.
• .Δ is the parameter set: It models all assumptions made on the context in which the
   DT model is used. .∀ λ ∈ Δ, dom(λ) is the set of all admissible values for .λ. .Δ is
   defined by variables which values can be updated by the context sensors pairing
   the DT with the ToI.
• . X is the input set: Its models the influences received from the DT environment
  .∀ x ∈ X, dom(x) is the set of all admissible values for x.
• .∑ is the semantic domain: This is the set of variables, which the phases are mapped
   onto. .∀ σ ∈ ∑, dom(σ ) is the set of all admissible values for .σ . .∑ is defined by
   variables which values can be updated by the system sensors pairing the DT with
   the ToI.
              j  j!    j!                                                         j    j!
• .Δ = {Δi? , Δi , Δi? , i ∈ Φ, j ∈ Φ} is the phase-to-phase transition set. .Δi? , Δi ,
           j!
   and .Δi? are respectively external, internal, and confluent transitions, from phase
   to phase (an internal transition occurs when the delay of the current phase elapses;
   an external transition is triggered by the receipt of an input; a confluent transition
   occurs when both internal and external transition conditions occur simultaneously),
   with:
          j
  – .Δi? =< ωi?, j , εi?, j , θi?, j > where:
          j!
  – .Δi =< εi, j! , ρi, j! , θi?, j > where :
          j!
  – .Δi? =< ωi?, j! , εi?, j! , ρi?, j! , θi?, j! > where:
      ·    ε
          . i?, j! ∈ ℘ (X ) is the condition of input receipt.
      ·    ε
          . i?, j! ∈ ℘ (∑) ∪ {N I L} is the condition of transition.
      ·   .ρi?, j! ∈ ℘ (∑) ∪ {N I L} is the output predicate.
      ·   .θi?, j! ∈ ℘ (∑) ∪ {N I L} is the action done in transition.
  – .Δ∗ =< i ∗ , θ ∗ , τ ∗ > is the initialization, giving the phase and action prior to the
     DT execution, where :
    As an example, let’s consider the DT of a smart city which operates a ring road
management system (Fig. 6). The DT collects data on weather and congestion through
road sensors and controls speed limit of the highway. Figure 4 illustrates the compo-
nents involved in the engineering of this DT (Fig. 7).
    The DT model of a ring road management system presents three phases: stopped
(i.e., accident or roadwork), regular (i.e., good weather conditions), and free-flowing
A Formal Framework for Digital Twin Modeling, Verification, and Validation      131
traffic. The quantity of vehicle (vehicle flow) will be noted F and the speed limit
(determined by the system) will be noted “S”. Roads’ number will be “R” and
number of vehicles “n”. Weather factor W), and reduced speed limit (i.e., due to bad
weather conditions or traffic jams).
   The formal specification of this model is illustrated by 8 and given by:
   DT Road Management = .< Δ, X, Y, Φ, ∑, Δ, Δ∗ >, where:
• .Δ = {n, W }, .n and .W can be updated with context data (number of vehicles and
   weather factor).
• . X = {in, war ning} and .Y = {F}, in can be updated with interface data (i.e.,
   modifying speed limit), and the speed limit set is sent through the output pin.
132                                                                     M. K. Traore et al.
• .∑ = {R, F}, R (number of roads) can be updated with system data (i.e., object
   fallen on a road), and . F (speed limit) is calculated, but the model is in regular or
   reduced phases.
• .Φ = {stopped, r egular, r educed}, with
        r egular
  – .Δstopped? =< in?on, N I L , N I L >, the external transition from stopped phase
     to regular phase arises when road manager switches it on, without any other
     condition (as indicated by the first NIL) or activity (as indicated by the second
     NIL).
       stopped
  – .Δr egular? =< in?o f f, N I L , N I L >, the external transition from regular phase
     to stopped phase arises when the road manager switches it off.
  – .Δrr egular?
         educed
                 =< war ning?o f f, N I L , I = W >, the internal transition from reg-
     ular phase to reduced phase arises when the weather factor (W) crosses a thresh-
     old.
       r egular
  – .Δr educed? =< war ning?on, N I L , I = W >, the internal transition from
     reduced to regular phase arises when the weather factor (W) crosses again the
     threshold.
• .Δ∗ =< r egular, N I L , 0 >, the speed limit is initially regular (before any simu-
   lation experiment).
   The adoption of this formal approach allows the DT developers to clarify the
intended concept behind each model that composes DT and allows possibility for
symbolic manipulation in order to formally check the consistency through several
verification and validation methods. The DEVs’ model we have just described pro-
vides a formal framework for structuring the “model” component (as depicted in
Fig. 2) of the Digital Twin. This formalism can be extended to rigorously describe
the various services that may interact with both the model and data layers. Adopt-
ing a consistent approach for specifying and understanding the intricate dynamics of
services within the Digital Twin ecosystem is essential for ensuring seamless interop-
erability and functionality. The utilization of such formalism will play a pivotal role
in achieving a comprehensive and well-defined representation of the entire Digital
Twin framework. This enables a more detailed analysis and validation of the intercon-
nected components, fostering a deeper understanding of the intricate relationships
between services, models, and data layers within the Digital Twin paradigm.
A Formal Framework for Digital Twin Modeling, Verification, and Validation           133
   However, the “DT Model Inference” layer is not explicitly addressed in the for-
malism proposed by Zeigler [37] (Fig. 5 left side). It remains an open challenge for
the scientific community to establish a method for formalizing the data inference over
model within the Digital Twin framework. A DEVS model can adeptly represent and
design various types of models, such as multi-agents, discrete events, process flows,
and cellular automata. Moreover, a DEVS model can also represent several services,
such as prediction, diagnosis, and optimization services. The scientific community
must now engage in defining a formalism that seamlessly accommodates the repre-
sentation of a data lake, regardless of its form (whether it be a database, compilation
of data sources, files, a centralized data entry point, ontology, etc.). This objective
necessitates a versatile approach that aligns with the diverse nature of data repre-
sentation in different Digital Twin applications, thereby ensuring a comprehensive
and adaptable framework for handling the intricacies of data management within the
Digital Twin paradigm.
The mechanisms of V&V for complex systems like Digital Twins can draw inspi-
ration from methods within the field of software engineering. It is well known that
software engineering uses several levels of tests to verify and validate any software
system [47]. These methods can be applied on each types of DT (DTP, DTI, and
DTA) dependent on the test level chosen. From a software engineering point of view,
V&V can be broken down into clearly identified stages that we propose to map
with DT architecture and described in Fig. 8: iterative unit tests (I), initially focus-
ing at the lowest level of testing. Here, each sub-components (including the three
main entities I.1-I.2-I.4 and inference model interaction I.3) are tested locally. The
goal of this testing level is to insure that the component being tested is conforming
to its specifications and ready to be integrated with other components of the DT.
The second level is integration testing (II of Fig. 8) which consists in insuring that
the interfaces between components are correct and the sub-components combined
to execute the DT’s functionality correctly. It goal is to assess the communication
between Twin of Interest (II.1) and the entire Digital Twin (II.2). Finally, the last
level is system testing, which is the process of verifying that the entire DT meets its
specified requirements (the all structure in Fig. 8).
Unit tests involve verifying and validating the smallest building blocks of the system.
The aim is to meet the requirements of each sub-component of the twin. In the frame-
work case, this involves addressing the data, models, and services’ sub-components
of Fig. 8. These unit tests are at the microscopic scale, thus corresponding to the tests
of DTI (Digital Twin Instance) type.
• Methods for Models’ verification and validation (I.1 of Fig. 8) are essential to
  ensure the accuracy and reliability of models within DTs. Different types of mod-
  els exist for different fields of application [48] (physical, mathematical, statistical,
A Formal Framework for Digital Twin Modeling, Verification, and Validation          135
  lies in the interactions between data and models. Here, the context of the Digital
  Twin imposes a very different treatment compared to other contexts. Thus, a data
  inference mechanism takes place in the DT. Data can, on the fly, influence the
  structures of the models [73]. It is therefore important to note that this part of the
  V&V opens up prospects for evolution in the scientific community.
• Integration testing between the ToI and data layer (II.1 of Fig. 8) is crucial to ensure
  seamless interaction and functionality. This integration testing phase focuses on
  verifying that the ToI effectively communicates with the data layer and that the
  data exchanged between them are accurate, consistent, and appropriately utilized.
  Testing scenarios may include validating data input/output mechanisms, ensuring
  compatibility between data formats and structures, and verifying the synchroniza-
  tion of data updates between the ToI and the Digital Twin’s data layer. Additionally,
  integration testing evaluates the responsiveness of the Digital Twin to queries and
  commands from the ToI, ensuring that the desired actions are executed correctly
  and in a timely manner. Through rigorous integration testing, developers can iden-
  tify and address any potential issues or discrepancies in the interaction between
  the Twin of Interest and the data layer, ultimately enhancing the reliability and
  effectiveness of the Digital Twin system.
• Entire Digital Twin (II.2 of Fig. 8) testing phase focuses on validating the inte-
  gration points between the three layers to ensure that they work harmoniously to
  achieve the intended objectives of the Digital Twin. Testing scenarios may include
  verifying the flow of data between the data layer and the model layer, ensuring
  that data inputs are accurately processed by the models and that the outputs are
  consistent with expected results. Additionally, integration testing evaluates the
  interaction between the model layer and the services layer, ensuring that services
  can effectively leverage the models to provide desired functionalities. Moreover,
  integration testing examines the integration points between the data layer and the
  services layer, validating that services can access and utilize data from the data layer
  efficiently. By thoroughly testing these integration points, developers can identify
  and address any potential issues or inconsistencies in the interactions between the
  layers of the Digital Twin, ultimately ensuring the reliability and effectiveness of
  the entire system.
System testing serves as a crucial phase aimed at evaluating the overall functionality,
performance, and behavior of the Digital Twin as a cohesive system. System testing
for a Digital Twin involves assessing its ability to fulfill its intended purpose and
objectives across various scenarios and use cases. This scale level can be associated
to Digital Twin Prototype. The comprehensive testing process examines the interac-
tion and integration of all components within the Digital Twin, including the data
layer, model layer, and services layer. It verifies the accuracy of data inputs, the effec-
tiveness of models in simulating real-world behavior, and the reliability of services
138                                                                  M. K. Traore et al.
in leveraging the Digital Twin for decision-making and optimization. System testing
also encompasses testing under different environmental conditions and stress levels
to ensure the robustness and resilience of the Digital Twin. By conducting system-
atic and rigorous system testing, developers can identify any potential weaknesses,
errors, or performance bottlenecks in the Digital Twin and take corrective actions to
enhance its overall quality and reliability.
Formal methods represent a powerful alternative for carrying out the V&V of Digital
Twin Instance. They provide structured and systematic techniques to ensure the
compliance of Digital Twins with the specifications and requirements of the real-
world system.
   Formal methods are approaches [75] based on mathematical and logical concepts
to specify, design, and verify software or hardware systems. They rely on formal
languages and formal analysis techniques to guarantee the correctness and safety
of systems. In the context of Digital Twins, formal methods can be used to specify
expected behaviors, detect design errors, and validate the consistency between the
virtual model and the real system.
   The first step in V&V of Digital Twins using formal methods is to develop a
precise formal modeling of the system. This modeling can be carried out using
formal languages such as Petri nets [76], automata [77], or formal specification
A Formal Framework for Digital Twin Modeling, Verification, and Validation           139
languages. By formally describing the system’s behavior and its interactions with the
environment, this modeling provides a solid foundation for analysis and verification.
   Once the formal model is established, various techniques of formal verification
can be used to ensure its correctness and compliance with specifications. Among
these techniques, model checking [78] involves automatically verifying if the model
satisfies certain specified properties by traversing the possibility tree to determine
the absence of deadlock or system termination. Theorem proving techniques [79], on
the other hand, use mathematical reasoning methods to demonstrate the correctness
of the model against a predefined set of properties.
   In addition to formal verification, formal validation of Digital Twins aims to ensure
that the virtual model accurately reflects the behavior of the real system. This step
may involve the use of formal simulation techniques [80], where the formal model
is executed in a simulation environment to evaluate its behavior under different
conditions. Comparing the results of formal simulation with real data validates the
accuracy and reliability of the Digital Twin.
   Formal methods offer a systematic and rigorous approach to the verification and
validation of systems, characteristics that can be beneficial for the Digital Twin
domain. By using formal languages, advanced analysis techniques, and specialized
tools, it is possible to ensure the reliability and accuracy of these virtual models,
paving the way for their use in critical applications such as industry, health care, and
transportation.
5 Conclusion
In this paper, our first contribution lies in proposing the design of an architecture
for understanding the Digital Twin, enabling high-level modeling. This architectural
understanding framework provides a structured approach to conceptualizing and
representing Digital Twins, facilitating their effective modeling and analysis through
three main levels: data level, model level, and service level.
    Building upon this foundation, our subsequent analysis delves into the question
of verification and validation of Digital Twins, a topic that has received limited atten-
tion in the existing literature despite the increasing prominence of this technology.
Recognizing this research gap, we have conducted an in-depth analysis to explore
the potential of existing methods in the state of the art to establish the conceptual
and methodological foundations for ensuring the accuracy and reliability of Digital
Twins. Our investigation has identified three primary axes: software engineering,
modeling and simulation, and formal methods, each offering valuable insights into
addressing the verification and validation challenges associated with Digital Twins.
This study serves as both a comprehensive review of the current state of research
and a roadmap for future endeavors in this emerging field. By consolidating existing
knowledge and pointing areas requiring further investigation, our work aims to facil-
itate significant progress in the verification and validation of Digital Twins, fostering
their widespread adoption across diverse application domains. Through collabora-
140                                                                               M. K. Traore et al.
tive efforts and continued research, we envision a future where Digital Twins play
an increasingly integral role in enhancing decision-making processes and driving
innovation across various industries.
References
21. Bolton RN et al (2018) Customer experience challenges: bringing together digital, physical
    and social realms. J Service Manage 29(5):776–808
22. Schroeder GN et al (2021) A methodology for digital twin modeling and deployment for
    industry 4.0. In: Proceedings of the IEEE, vol 109, issue 4, pp 556–567. https://doi.org/10.
    1109/JPROC.2020.3032444.
23. White G et al (2021) A digital twin smart city for citizen feedback. Cities 110:103064
24. Barnabas J, Raj P (2020) The human body: a digital twin of the cyber physical systems. Adv
    Comput 117(1):219–246
25. Bailenson JN, Segovia KY (2010) Virtual doppelgangers: psychological effects of avatars who
    ignore their owners. In: Online worlds: convergence of the real and the virtual, pp 175–186
26. Jain P et al (2019) A digital twin approach for fault diagnosis in distributed photovoltaic
    systems. IEEE Trans Power Electronics 35(1):940–956
27. Zhou X et al (2023) A fuzzy-set-based joint distribution adaptation method for regression and
    its application to online damage quantification for structural digital twin. Mech Syst Signal
    Process 191:110164
28. Mama D, Kaba TM (2023) Formal approach to digital twin specification. In: Annual Modeling
    and Simulation Conference (ANNSIM). IEEE, pp 233–244
29. Barricelli BR, Casiraghi E, Fogli D (2019) A survey on digital twin: definitions, characteristics,
    applications, and design implications. IEEE Access 7:167653–167671
30. David J et al (2020) Characterising the digital twin: a systematic literature review. CIRP J
    Manuf Sci Technol 29:36–52
31. VanDerHorn E, Mahadevan S (2021) Digital twin: generalization, characterization and imple-
    mentation. Decis Support Syst 145:113524
32. Mengnan L et al (2021) Review of digital twin about concepts, technologies, and industrial
    applications. J Manuf Syst 58:346–361
33. Tao F et al (2022) Digital twin modeling. J Manuf Syst 64:372–389. ISSN: 0278-
    6125. https://doi.org/10.1016/j.jmsy.2022.06.015. https://www.sciencedirect.com/science/
    article/pii/S0278612522001108
34. Tao F et al (2021) Theory of digital twin modeling and its application. Comput Integr Manuf
    Syst 27:1–15
35. Thelen A et al (2022) A comprehensive review of digital twin–part 1: modeling and twinning
    enabling technologies. Struct Multidisc Optimization 65(12):354
36. Cronrath C, Ekström L, Lennartson B (2020) Formal properties of the digital twin—
    implications for learning, optimization, and control. In: 2020 IEEE 16th international Con-
    ference on Automation Science and Engineering (CASE). IEEE, pp 679–684
37. Zeigler BP (1976) The hierarchy of system specifications and the problem of structural infer-
    ence. In: PSA: proceedings of the biennial meeting of the philosophy of science association,
    vol 1976, 1. Cambridge University Press, pp 226–239
38. Worden K et al (2020) On digital twins, mirrors, and virtualizations: frameworks for model
    verification and validation. In: ASCE-ASME J Risk Uncertainty Eng Syst Part B: Mech Eng
    6(3):030902
39. Zheng X et al (2017) Perceptions on the state of the art in verification and validation in
    cyber-physical systems. IEEE Syst J 11(4):2614–2627. https://doi.org/10.1109/JSYST.2015.
    2496293
40. Khan A et al (2018) Digital Twin for legacy systems: simulation model testing and validation.
    In: 2018 IEEE 14th international Conference on Automation Science and Engineering (CASE),
    pp 421–426. https://doi.org/10.1109/COASE.2018.8560338.
41. Hasan A et al (2023) Predictive digital twins for autonomous surface vessels. In: Ocean
    engineering, vol 288, p 116046. ISSN: 0029-8018. https://doi.org/10.1016/j.oceaneng.2023.
    116046. https://www.sciencedirect.com/science/article/pii/S0029801823024307. A formal
    framework for digital twin modeling, verification and validation 25
42. Argota Sánchez-Vaquerizo J (2022) Getting real: the challenge of building and validating
    a large-scale digital twin of Barcelona & rsquo;s traffic with empirical data. ISPRS Int J
    Geo-Information 11(1). ISSN: 2220-9964. https://doi.org/10.3390/ijgi11010024. https://www.
    mdpi.com/2220-9964/11/1/24
142                                                                              M. K. Traore et al.
43. Zheng X, Julien C (2015) Verification and validation in cyber physical systems: research chal-
    lenges and a way forward. In: 2015 IEEE/ACM 1st international workshop on software engi-
    neering for smart cyber-physical systems, pp 15–18. https://doi.org/10.1109/SEsCPS.2015.
    11
44. Wallace DR, Fujii RU (1989) Software verification and validation: an overview. IEEE Softw
    6(3):10–17
45. Traore MK, Gorecki S, Ducq Y (2022) A simulation based approach to digital twin’s inter-
    operability verification & validation. In: “Workshop interoperability challenges and solutions
    within industrial networks” co-located with 11th international conference on Interoperability
    for Enterprise Systems and Applications (I-ESA 2022), vol 3214
46. Gore R, Diallo S (2013) The need for usable formal methods in verification and validation. In:
    2013 Winter Simulations Conference (WSC), pp 1257–1268. https://doi.org/10.1109/WSC.
    2013.6721513
47. Collofello JS (1988) Introduction to software verification and validation. Carnegie Mellon
    University, Software Engineering Institute
48. Fishwick PA, Miller JA (2004) Ontologies for modeling and simulation: issues and approaches.
    In: Proceedings of the 2004 Winter Simulation Conference, vol 1. IEEE
49. Thacker BH et al (2004) Concepts of model verification and validation
50. Wang C et al (2018) Detecting aberrant behavior and item preknowledge: a comparison of
    mixture modeling method and residual method. J Educ Behav Stat 43(4):469–501
51. Rasmus B et al (2008) Cross-validation of component models: a critical look at current methods.
    Analy Bioanalytical Chem 390:1241–1251
52. Darlington RB, Hayes AF (2016) Regression analysis and linear models: concepts, applications,
    and implementation. Guilford Publications
53. Yue J (2009) Variance analysis in software fault prediction models. In: 20th international
    symposium on software reliability engineering. IEEE, pp 99–108
54. Koch K-R (2013) Parameter estimation and hypothesis testing in linear models. Springer Sci-
    ence & Business Media
55. Kim WC, Ahn SC, Kwon WH (1995) Stability analysis and stabilization of fuzzy state space
    models. Fuzzy Sets Syst 71(1):131–142. 26 Mamadou Kaba Traore Simon Gorecki and Yves
    Ducq
56. Gidskehaug L, Anderssen E, Alsberg BK (2008) Cross model validation and optimisation of
    bilinear regression models. Chemometrics Intell Lab Syst 93(1):1–10
57. Confalonieri R et al (2009) Multi-metric evaluation of the models WARM, CropSyst, and
    WOFOST for rice. Ecol Modelling 220(11):1395–1410
58. Willmott CJ, Matsuura K (2005) Advantages of the mean absolute error (MAE) over the root
    mean square error (RMSE) in assessing average model performance. Climate Res 30(1):79–82
59. Murat K, Arjan D (2017) Quality of service (QoS) in software defined networking (SDN): a
    survey. J Netw Comput Appl 80:200–218
60. Data Warehousing Institute. Data quality and the bottom line: achieving business suc-
    cess through a commitment to high quality data. http://www.dw-institute.com/ (visited on
    02/08/2024)
61. Chengalur-Smith ISN, Ballou DP, Pazer HL (1999) The impact of data quality information on
    decision making: an exploratory analysis. IEEE Trans Knowl Data Eng 11(6):853–864
62. Monica S (2006) Data quality: concepts. Data-Centric Systems and Applications, Springer,
    Methodologies and Techniques
63. Office of Management and Budget. Information quality guidelines for ensuring and maximizing
    the quality, objectivity, utility, and integrity of information disseminated by agencies. http://
    www.whitehouse.gov/omb/fedreg/reproducible.html (visited on 01/17/2006)
64. DQI (2004) Data quality initiative framework. Project report. https://nces.ed.gov/fcsm/pdf/
    FCSM.20.04_A_Framework_for_Data_Quality.pdf (visited on 2004)
65. Batini C et al (2009) Methodologies for data quality assessment and improvement. ACM
    Comput Surv (CSUR) 41(3):1–52
A Formal Framework for Digital Twin Modeling, Verification, and Validation                    143
Abstract Digital Twins are becoming more prevalent in a wide range of industries
such as manufacturing, construction, smart city, and healthcare for various purposes,
including observing, predicting, optimizing, and controlling. Digital Twins are in
the early adoption stage. Currently, few standards directly address Digital Twins
and a commercial ecosystem of Digital Twins has not been well established. Devel-
oping and implementing Digital Twins present significant challenges. Most current
Digital Twin applications are customized solutions, which are expensive to create
and difficult to integrate with other systems. Foundational work is needed to support
an open marketplace for Digital Twin developers, users, and technology service
providers. This includes the development of standardized frameworks, reference
models, and interfaces to provide a solid foundation for ensuring interoperability,
reliability, validity, security, and trust. This Chapter identifies implementation chal-
lenges for Digital Twins for manufacturing, reviews relevant standardization efforts,
introduces the ISO Digital Twin framework standard for manufacturing, presents use
cases, and discusses potential research topics and future standardization directions.
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2024             145
M. Grieves and E. Hua (eds.), Digital Twins, Simulation, and the Metaverse, Simulation
Foundations, Methods and Applications, https://doi.org/10.1007/978-3-031-69107-2_7
146                                                                       G. Shao et al.
1 Introduction
proposed [4, 5]. Many Digital Twins today are implemented for real-time state moni-
toring. The aerospace industry has used Digital Twins to monitor jet engines for many
years. These types of Digital Twins use one-way synchronization. They receive data
from the physical object but do not provide control feedback.
    Galli et al. [6] analyzed a variety of proposed frameworks for building Digital
Twins. This research focused on the structure of the frameworks to find correlations
in terms of form and conceptualization. The results showed three types of archi-
tectures used for the Digital Twin. The first is the traditional and is based on the
original framework proposed by Grieves and Vickers [7]. This architecture supports
the Digital Twin in paralleling the real system, discussing interfaces and interoper-
ability with operations management systems. It also ensures synchronization between
the physical system and its corresponding digital twin. The second is the “service-
oriented” type, which consists of four components: a physical shop floor, a virtual
model of the shop floor, a service system, and the shop floor Digital Twin data. The
third architecture is the fractal, where local or specialized Digital Twins make up
the global Digital Twin, each with a similar structure. Khan et al. proposed a six-
dimensional framework to include (1) a physical asset, (2) a digital duplicate of the
physical asset, (3) data generated by the physical asset, (4) programs that improve
product performance and make the production process more efficient, (5) spiral-rings
like iterations that generate a more optimized product or efficient production process,
and (6) synchronization between the physical asset and the digital twin [4].
    Despite these efforts, the Digital Twin technology is still in its early stages. There
is a lack of universal definitions, implementation frameworks, and protocols. There is
also a lack of comprehensive and in-depth analysis of Digital Twins from the perspec-
tive of concepts, technologies, and industrial applications research [8]. Further, the
ecosystem of Digital Twins is not well established, and most Digital Twins are devel-
oped using customized solutions. Customized solutions do not support reuse and are
costly and time-consuming. A systematic approach is needed to characterize and
manage Digital Twin subsystems to ensure cross-disciplinary interoperability and
credibility of the Digital Twin. Standards are the solutions to enable such interop-
erability. However, only a few standards have been developed for Digital Twins so
far.
    According to Accenture research, companies are not taking full advantage of
Digital Twins because most Digital Twin applications are standalone for single func-
tions, which focus on functional optimization instead of enterprise optimization, and
have no comprehensive strategy for data integration and sharing [9]. Additionally, if
a Digital Twin is the current representation of a physical asset, the associated Digital
Twin data at any life cycle stage could be helpful for future asset management.
The architecture or mechanism supporting the flow of information about a product’s
performance and use from design, production, use, disposal, and recycling is referred
to as the digital thread [10]. Using the digital thread will enable the traceability of
Digital Twins from requirements to the retirement of the physical asset. Interoper-
ability among Digital Twins for different life cycle stages through digital threads
should help overcome these challenges. Standardized data representation will help
avoid customized Digital Twin development and the duplication of efforts. Digital
148                                                                         G. Shao et al.
threads can also provide an integrated view of the physical asset for Digital Twin
development, avoiding redundancy during information exchange.
    To make the best use of Digital Twins, manufacturers need to apply interoperability
standards from both systems of systems and life cycle perspectives. In addition,
standards on vocabulary, reference architecture, and trustworthiness can help ensure
the interoperability, value, and credibility of the Digital Twins. These standards
will enable manufacturers to build, manage, and deploy their Digital Twins more
efficiently.
    This Chapter focuses on a standardized approach to building Digital Twins for
advanced manufacturing. It identifies current challenges for manufacturers to imple-
ment their Digital Twins, reviews relevant standards efforts, introduces the ISO
Digital Twin framework standard, presents use-case scenarios, and discusses some
potential standards development research directions. The rest of the chapter is orga-
nized as follows: Sect. 2 discusses various applications of Digital Twins, Sect. 3
presents how standards help address the challenges of implementing and adopting
Digital Twins that manufacturers face, Sect. 4 introduces the ISO standard, ISO
23247—Digital Twin Framework for Manufacturing, Sect. 5 discusses additional
relevant standards, Sect. 6 describes a case study of applying standards for building
a Digital Twin of a robot work cell, Sect. 7 discusses potential topics and research
directions toward standardization, and Sect. 8 summarizes the chapter and presents
future work.
. Diagnostic Digital Twins analyze what has gone wrong and why it has happened
  or is happening to their physical counterparts. Diagnosis includes analyses of the
  impact of a data input and an operational strategy on key performance indicators
  (KPIs). For example, a product’s cycle time increase may be caused by machine
  breakdowns or bad scheduling decisions. Diagnostics are supported by technolo-
  gies and methodologies such as simulation, data mining, machine learning, and
  data analytics.
. Predictive Digital Twins predict what will happen and when it will happen. These
  kinds of Digital Twins can be used to estimate when a machine tool or robot
  deterioration will likely reach a point of failure based on past performance patterns.
  They can also pinpoint the cause or source of failure. Enabling technologies may
  include modeling and simulation, predictive data mining, and parameter tracking.
  For example, Digital Twins integrated with machine learning can be used to predict
  cycle times for incoming manufacturing orders [5].
. Prescriptive Digital Twins provide the influence and control of the physical coun-
  terparts and decide how to make it happen based on the objectives of the Digital
  Twins. These kinds of Digital Twins can help identify the strategies and inputs
  leading to optimal performance. For example, prescriptive analytics provide the
  best possible input parameters and methods that enable cycle-time reduction and
  increase throughput. Enabling technologies may include simulation, optimization,
  and control.
. Intelligent Digital Twins are envisaged to control their physical counterparts based
  on the strategies and parameters identified by the prescriptive Digital Twins. These
  kinds of Digital Twins can learn new strategies based on the data collected and
  take actions accordingly, which could include dynamically adjusting themselves
  based on the changes in their physical counterparts to keep up with them or based
150                                                                        G. Shao et al.
   on the Digital Twin objectives to ensure the physical counterparts operate opti-
   mally. Artificial Intelligence (AI), including large language models, is the primary
   modeling technique used for modeling the intelligent Digital Twin. Other enabling
   technologies may include simulation, optimization, automation, and control.
   The kind of Digital Twin to be implemented depends on the use case and is driven
by the objective and scope of the Digital Twin.
As discussed in Sect. 1, there are still significant challenges for manufacturers, espe-
cially Small and Medium-sized Enterprises (SMEs), to implement their Digital Twin
applications efficiently and effectively. Current implementations mostly use ad hoc,
customized approaches. Customized solutions not only increase the development
time and cost but also make it challenging to integrate with other systems and do not
support reuse. Standards are needed for manufacturers to go beyond custom, expen-
sive Digital Twins to an affordable marketplace of products and tools for Digital
Twins. Standards, such as frameworks, reference models, and interfaces, will provide
a solid foundation for Digital Twin developers, users, and technology and service
providers to ensure interoperability, reliability, validity, security, and trust.
   Standards for Digital Twins will facilitate the composition and integration of
Digital Twins by providing guidelines, methodologies, common terminologies, archi-
tectures, and interface specifications. These standards can help make the creation,
integration, update, and validation of Digital Twins more accurate and consistent.
Standards can also help formalize requirements for Digital Twin projects, enable the
use of building blocks for Digital Twin implementations, analyze Digital Twin perfor-
mance, communicate between suppliers, partners, and customers, secure Digital
Twin information and protect privacy, and facilitate the verification and validation
of Digital Twins according to stakeholders’ requirements. Ultimately, standards will
help achieve “plug and play,” i.e., enabling interoperability between Digital Twins
and among software and hardware from various vendors. In different areas, standards
support the following:
. Defining a common language for data representation, communication protocols,
  and Application Programming Interfaces (APIs) to ensure Digital Twin systems
  can understand and interact with each other’s data. Examples of APIs include
  those for data query, data update, and data synchronization.
. Developing shared metadata and ontologies to describe the properties, attributes,
  and relationships within Digital Twins for easy mapping and translating data
  between Digital Twins.
Digital Twins for Advanced Manufacturing: The Standardized Approach                  151
. Developing tools or middleware that can map data from one Digital Twin’s format
  to another using techniques such as data transformation, data normalization, or
  data translation.
. Implementing robust security measures to ensure only authorized Digital Twins
  can access and interact with each other using techniques such as authentication,
  encryption, and access control.
The ISO standard, ISO 23247—Digital Twin Framework for Manufacturing, was
created to facilitate the implementation of Digital Twins in manufacturing. The stan-
dard defines a “Digital Twin in Manufacturing” as a “fit for purpose digital repre-
sentation of an observable manufacturing element with synchronization between
the element and its digital representation [12].” It provides a generic guideline,
a reference architecture, and a framework for Digital Twin applications in manu-
facturing. The standard also provides examples of data collection, data commu-
nication, integration, modeling, and applications of relevant standards [12]. The
standard provides procedures for manufacturers and solution providers to analyze
Digital Twin requirements, define scope and objectives, use common terminologies,
comply with a generic reference architecture, and integrate multiple existing stan-
dards for various purposes. The framework includes the sub-entities and components
as building blocks for manufacturers to pick and choose for their own case-specific
Digital Twin development. It helps manufacturers systematically identify and deter-
mine subsystems and components, their relationships, and the characteristics of their
interactions from which appropriate standards can be selected for interoperability.
   A key feature of ISO 23247 is that it enables the deployment of the digital thread,
implying that model-based engineering standards for various stages of a product life
cycle can be included in the framework. For example, for the Digital Twins devel-
oped to support a product at different stages, including design, manufacturing, and
inspection, relevant standards such as Standard for the Exchange of Product Model
Data (STEP) [13], MTConnect [14], and Quality Information Framework (QIF) [15]
can be applied. Therefore, the standard supports Digital Twins’ compatibility and
interoperability throughout the life cycle stages, allowing information reuse and
traceability.
   The standard series includes four parts: (1) overview and general principles, (2)
reference architecture, (3) digital representation, and (4) information exchange. The
reference architecture in the standard includes a reference model with domains and
entities. There are four domains (layers), each with a logical set of tasks and functions
performed by functional entities. Figure 2 shows the entity-based reference model
and an illustration of the four domains and their interactions [16]. Each domain is
briefed as follows:
152                                                                            G. Shao et al.
Fig. 2 Functional view of the digital twin reference model for manufacturing
   The cross-system entity in Fig. 2 is an entity that resides across domains to provide
common functionalities such as data translation, data assurance, and security support.
Digital Twins can be developed based on the Digital Twin Framework depicted within
the dotted line in Fig. 2. The framework supports the applications of IoT infrastructure
for data collection, communication protocols for data transmission, and information
flows between entities of different domains—OMEs, Data Collection and Device
Control, Digital Twin Core, and User layers.
Digital Twins for Advanced Manufacturing: The Standardized Approach              153
   With these relevant standards from various functional categories, users can select
those applicable to their Digital Twin implementations.
In this section, a robotic work cell Digital Twin is discussed to exemplify the appli-
cation of relevant standards. The work cell consists of collaborative robot arms for
material handling and machine tending, a CNC machine tool for machining, and
a Coordinate Measuring Machine (CMM) for product geometry measurements and
quality control. Figure 3 shows the workflow and equipment for the use-case scenario
in the work cell. The workflow includes receiving parts, loading parts to the CNC
(by ROBOT #1), cutting the parts, unloading parts from the CNC (by ROBOT #1),
loading parts to the CMM (by ROBOT #2), inspecting the parts, and offloading
parts from the CMM (by ROBOT #2). Parts that fail the inspection will be sent to
the rework buffer by ROBOT#2. The cell has a single input location and a single
output location. The robots are fitted with a 2F-85 gripper to handle components
with different geometries effectively.
   One of the research objectives is to build a process digital twin that is a composite
of the Digital Twins of each individual component.
Fig. 3 Workflow through machines and equipment in the robot work cell
158                                                                       G. Shao et al.
   The standards used in implementing the use-case include ISO 23247, MTCon-
nect, STEP, QIF, and ASME Verification, Validation, and Uncertainty Quantification
(VVUQ).
Based on the Digital Twin framework introduced in Sect. 3, each piece of equipment
in the work cell is treated as an OME, for which data need to be collected, and Digital
Twins for various scenarios must be developed. In this section, we focus on one robot
arm in the work cell to showcase the method of Digital Twin development.
    The ISO Digital Twin framework standard is instantiated for the robot arm (UR5e),
whose operations include picking and loading a workpiece to a CNC machine tool.
Figure 4 illustrates the instantiation of the framework for the robot arm. The figure
shows data being collected from the robot arm through a MTConnect adapter. The
Digital Twin entity comprises the simulation model of the robot arm and analytical
models that manipulate real-time data to support decision-making. The modeling
method and environment support the three-dimensional geometry of the robot arm
components, including the robot base, links, joints, end effector, and workpiece. The
user entity includes the developer and user of the Digital Twin, production software
systems, or other Digital Twins. Similarly, the standard can be instantiated for the
cutting part, the CNC machine, and the CMM.
6.2 MTConnect
The process of acquiring data and building a scalable data pipeline for the work cell
involves multiple activities: (1) collecting real-time operational data from the robot,
(2) leveraging the MTConnect standard to provide defined machine data and make it
available in a standard format, and (3) developing a set of tools to enable client-side
use of the MTConnect agent.
Operational Data Collection: To collect the operational data from the UR5e robot
arm, we utilized the Universal Robot’s Real-Time Data Exchange (UR-RTDE) inter-
face with APIs supporting data collection. Several data items collected include
angular position, velocity, acceleration, torque, current, and temperature for each
of the six joints of the robot arm. Universal Robots also shows the data items (and
the corresponding units) that the vendor offers through the UR-RTDE interface.
MTConnect Interface: Implementing the MTConnect standard requires an adapter
and an agent. The adapter serves as a data collection element from the equipment
while the agent collects data from the adapter. Many machine vendors provide a
preinstalled adapter but not the UR5e robot. The adapter was installed on an interface
connected to the robot arm. The adapter packages data into a format that is readable
Digital Twins for Advanced Manufacturing: The Standardized Approach                        159
Fig. 4 Implementation method of building a Digital Twin for the robot based on ISO 23247
by the agent. The agent provides an application interface to retrieve the MTConnect
data gathered from the adapter. Figure 5 shows the data flow from a physical device
(UR5e) to the Digital Twin. A semantic structure was provided for the physical
data generated by the UR5e robot arm through MTConnect. This semantic structure
includes data tags and units based on the MTConnect 2.0 Standard [42]. Using the
Python UR-RTDE API, a socket-based adapter that sends MTConnect-compliant
data to the agent was developed. An instance of the MTConnect agent receives and
serves the data in a machine-readable format.
Client-Side Integration: The MTConnect agent is implemented in C++ and displays
data in XML format on a Hypertext Transfer Protocol (HTTP) server. The tool
used for modeling the robot work cell is the Simulink/Simscape, which can input
comma-separated values (CSV) where multiple data items over a time interval are
Fig. 6 The physical and virtual model of the UR5e robot arm with attached gripper
recorded and synced up to their respective timestamps (UTC format). The tools have
been developed to parse the XML output from the MTConnect Agent, populate a
2-dimensional array, and store the array in a CSV file.
6.3 Step
To develop a physical model of the robot arm, Computer Aided Design (CAD) models
of its components and those of the work environment are fundamental. These models
include the description of the geometry of the links and how they are connected to
the robot arm. These CAD models are imported into the Digital Twin environment to
create a physical model of the UR5e robot arm. The CAD models for the gripper are
“assembled” into the end effector in the AutoCAD Inventor environment, exported
to the Digital Twin environment, and “attached” to the robot arm model. Figure 6
shows the physical and virtual models of the robot arm with the attached end effector.
    The CAD models of the part, CNC, and CMM can be in STEP standard format,
which allows easy exchange and visualization. The design part model in STEP will
also be used to compare the measurements of the finished product.
QIF provides an integrated model for manufacturing quality information. The CMM
provides measurement information about a product regarding conformance to spec-
ifications per Geometric Dimensioning and Tolerancing (GD&T) design. Design
tolerances are defined by the amount a feature is allowed to vary from the nominal.
Digital Twins for Advanced Manufacturing: The Standardized Approach                  161
Digital Twins are complex systems that are sometimes used as virtual testbeds for
verification and validation of systems, especially in cases where actual tests are
complex to perform. However, the quality of decisions made with Digital Twins
depends on the validity of the underlying models. A valid Digital Twin should accu-
rately describe the system that changes over time. Thus, the development of the
Digital Twins needs to be validated before its use in supporting decision-making.
Verification, Validation, and Uncertainty Quantification (VVUQ) standards need to
be followed to ensure that the Digital Twin is built correctly and that the right Digital
Twin has been built. Zhang et al. [45] discuss verification and validation methods
for Digital Twins, which are categorized into qualitative, quantitative, and integrated
methods. Both qualitative and quantitative methods require metrics for Digital Twin
validation. These metrics include credibility/fidelity, complexity, standardization,
and capability maturity of model construction.
    Hua et al. [46] summarized general strategies to validate a Digital Twin. These
include visual inspection of the Digital Twin for correctness using established stan-
dards, testing properties of the Digital Twin, model-based testing using methods such
as input–output conformance testing, and machine learning or artificial intelligence-
based testing. Kibira and Weiss [47] used a model-based approach to validate the
Digital Twin model of a robot arm. Joint position and orientation data, velocity data,
and acceleration data were collected from the physical twin for validation under the
model-based approach.
    Verification and validation standards include:
. ASME V&V 10: Standard for Verification and Validation in Computational Solid
  Mechanics [48].
. ASME V&V 20: Standard for Verification and Validation in Computational Fluid
  Dynamics and Heat Transfer [49].
. ASME V&V 40: Assessing Credibility of Computational Modeling Through
  Verification and Validation: Application to Medical Devices [50].
. Other ASME V&V 50, 60, 70, and 80 standards are under development. V&V
  50 standards are for advanced manufacturing, and V&V 70 standards are for
  data-driven models.
162                                                                         G. Shao et al.
The current four parts of the ISO 23247 series provide a fundamental generic Digital
Twin framework for manufacturing. The framework can be extended to industries
that employ specialized manufacturing processes and technologies. Future work on
this standard may result in new additions supporting the development and validation
of Digital Twins. The new research topics for extension of the standard include (1)
digital thread for Digital Twins, (2) Digital Twin composition, (3) ontologies of the
Digital Twin framework to clarify the entities and relationships, (4) building Digital
Twins from reusable components to increase the consistency and reduce the develop-
ment time, (5) credibility assessment of Digital Twins to increase the trustworthiness
and value for decision-making, (6) Digital Twins and the metaverse to provide guide-
lines that enable the integration between Digital Twins and industrial metaverse, (7)
plug and play Digital Twin integration by standardizing interfaces with customers’
environment and application platforms, and (8) extending the framework to specific
sectors, e.g., semiconductor manufacturing, biomanufacturing, and additive manu-
facturing to address domain-specific needs [51]. The following subsections discuss
the potential new parts of the standard that could enable better and easier Digital
Twin development.
Digital Twin composition implies that multiple Digital Twins are developed and
integrated with the support of a digital thread. For example, the Digital Twin of a
part and that of the machine that manufactures the part can interact dynamically
and seamlessly. When the Digital Twin of a cutting tool, a machine tool, and a part
interact, they can be used to determine the tool wear, tolerance conformance, and the
machine’s health. Digital Twins of multiple partners coordinate and communicate in
real time in a supply chain. However, it is challenging to aggregate, compose, and
integrate multiple Digital Twins and applications to achieve a new goal. Standard-
based methods and guidelines will help achieve this, reduce development time, and
mitigate risks for such undertakings.
   This part of ISO 23247 will provide guidelines on enabling multiple Digital Twins
to communicate and interoperate effectively. The new part could provide generic
methodologies, principles, and examples to help users understand the purpose of
the Digital Twin and develop the appropriate Digital Twin to address the identified
problem(s). Relevant standards and technologies could be selected and applied to
demonstrate the integration.
   This part of ISO 23247 will also specify Digital Twin composition by defining
principles, showing methodologies, and providing use-case examples of Digital
Twin configuration, communication, aggregation, composition, integration, and
collaboration during manufacturing.
The current four parts of the ISO 23247 standard define the terms, relationships,
components, and processes necessary for developing a Digital Twin and provide
guidelines for Digital Twin implementation. However, logical formalism does not
support it, which may lead to inconsistent implementation. The ontologies for the
Digital Twin framework will enable the definitions of terms that are both human-
understandable and computer-processable, which result in an unambiguous repre-
sentation of a particular construct and consistent interpretation, regardless of the
initial data source. It also enables explicit representation of the connections between
different terms; different connections permit a consistent presence and representa-
tion of the required metadata. A potential new part of ISO 23247 on this topic could
provide an ontology for the Digital Twin framework.
164                                                                         G. Shao et al.
Digital Twins could be developed for different control levels depending on the
application, including equipment, work cells, production lines, factories, and supply
chains. While some approaches exist to support model component reuse [52], most
of them are not explicitly designed for Digital Twins. Therefore, almost all Digital
Twins are built from scratch, which makes implementations time-consuming and
costly. Customized designs also make a Digital Twin challenging to modify, extend,
and reuse. Manufacturing knowledge, information attributes, and use-case configu-
rations are often developed using different specialized abstractions for each appli-
cation. The reusability of Digital Twin components in a Digital Twin library could
considerably reduce the development cost, time, and the required level of expertise.
    A potential new part of ISO 23247 on this topic could provide guidelines on
building component libraries and creating templates for organizing data, informa-
tion, and models. Reusable Digital Twin components may include templates for
data collection, common information attributes, and modular models. Digital Twin
development would be supported by enabling technologies and relevant standards for
various Digital Twin functions. The new part could provide generic methodologies,
architectures, frameworks, knowledge bases, and examples for building and using
Digital Twin component libraries.
The current four parts of ISO 23247 do not cover VVUQ and testing on Digital Twins.
Given the potential use of Digital Twins in critical decision-making for various manu-
facturing applications, the results generated by Digital Twins must be trustworthy
for real manufacturing needs. Model credibility assessment, including VVUQ tech-
niques, must be applied throughout the life cycle of Digital Twins. VVUQ should be
embedded in the design, creation, and deployment of Digital Twins to establish trust
in the model and its outputs [53]. Verification and Validation (V&V) activities are
necessary to ensure that a Digital Twin meets its intended purpose and design goals.
Uncertainty Quantification (UQ) produces a measure of performance that users can
apply as part of a credibility assessment for a given Digital Twin. VVUQ for Digital
Twins should be a continual process that adapts to changes in the OME and its digital
representation, data inputs, and decisions made [53]. The credibility assessment of
Digital Twins may also include factors beyond VVUQ.
   Digital Twin testing needs a test system comprising a set of tests for both the
OME and its Digital Twin. The test system should also define what an acceptably
valid Digital Twin should look like. Grieves proposed a virtual testing method for
manufactured products, which can be adapted to Digital Twins [54]. For example,
suppose the test system can run a set of tests, and the results of the Digital Twin can’t
be distinguished from those of the OME within a predefined probability threshold. In
Digital Twins for Advanced Manufacturing: The Standardized Approach                165
that case, the Digital Twin can be regarded as a reasonable representation of the OME.
Trust in a Digital Twin also involves trust in the data collected from the OME, the
model used in the Digital Twin, the data updating procedure, and the recommended
decisions. All these aspects should have a measurable uncertainty, whose existence
means that validation (comparison with reality) needs to be treated as a statistical
process. Comparison of actual data with model results can be used to estimate the
probability that the Digital Twin is a consistent representation of the OME.
   Currently, there is no standard process for reporting VVUQ for digital twins.
Developing robust VVUQ processes for digital twins remains a challenge. A potential
new part of ISO 23247 on this topic could provide guidelines on and methodologies
for how to measure uncertainty, how to perform VVUQ and testing for Digital Twins,
how to select or construct a credibility assessment framework that supports these
activities, and how to assess the credibility of the developed Digital Twins.
The metaverse can support monitoring the manufacturing system in real time, both
visually and from a metric standpoint. It could also provide users with an immersive
experience. This is now possible because of the maturity of technologies for virtual
reality (VR), augmented reality (AR), and extended reality (XR), which can enhance
users’ visualization experience for manufacturing. For example, it has been demon-
strated that AR technologies can be integrated with three-dimensional geometrical
product specification and verification standards and practices [55].
   A significant feature of the metaverse is the immersive visualization experience
along with its human–machine interface. The hardware and software technologies
developed for the metaverse can be used by the Digital Twin framework for manufac-
turing, especially in cases where there is human involvement. For example, the user
domain (shown in Fig. 2) and user entity of the ISO 23247 standard can use human–
machine interfaces provided by the metaverse. Alternatively, a metaverse may be a
parallel virtual world that may subsume some of the Digital Twins of a manufac-
turing enterprise that the metaverse represents. A new part of the ISO 23247 series
could include the metaverse concept, its definition, possible scenarios for integrating
with manufacturing Digital Twins, including human Digital Twins, guidelines, and
methodologies for such integration.
Based on the generic framework provided by the initial four parts of the ISO 23247
series, extensions can be developed as new parts of the standard for specific manufac-
turing sectors such as biomanufacturing, semiconductor manufacturing, and additive
166                                                                        G. Shao et al.
manufacturing. The new parts may include specialization of the Digital Twin frame-
work by adding new functional entities or modifying existing functional entities to
fit the new requirements. The new parts may also have the use cases implemented
for those manufacturing sectors. These use-case implementations may, in turn, help
identify new standardization requirements for that manufacturing sector.
    The emerging biomanufacturing sector can use the generic framework to develop
its Digital Twins, which may constitute a new part of the ISO 23247 series. Similarly,
Digital Twins for additive manufacturing may have substantial potential to improve
process control, and a new part in the ISO 23247 series can be dedicated to additive
manufacturing. In semiconductor manufacturing, an extension of the standard can be
developed to address challenges such as obtaining datasets for constructing Digital
Twin models and cybersecurity associated with the Digital Twin. Other standards
development organizations may adopt the current ISO 23247 series to create Digital
Twins for their customer industries.
8 Summary
Digital Twins are becoming more prevalent in a wide variety of industries, including
manufacturing. However, a standards-based ecosystem of Digital Twins has not yet
been established. Developing and integrating Digital Twins presents significant chal-
lenges. Foundational work is needed to support an open marketplace for Digital Twin
developers, users, and technology and service providers. This includes the devel-
opment of standardized frameworks, reference models, interface specifications, and
VVUQ methodologies to provide a solid foundation for ensuring the interoperability,
validity, security, and trust of Digital Twins. This Chapter focuses on applying Digital
Twins in manufacturing within a framework of standards, identifies current chal-
lenges, reviews relevant standardization efforts, and introduces the ISO Digital Twin
framework standard for manufacturing, ISO 23247. This work also discusses some
potential research directions and future standardized topics. A use case is presented
to illustrate the Digital Twin development process by applying relevant standards.
The example is obtained from the Digital Twin development efforts for a robot work
cell in a Digital Twin Lab at NIST.
    Future efforts include performing measurement science research to support the
development and integration of Digital Twins in manufacturing, working with Indus-
trial consortia and standards development organizations to prioritize the standard-
ization topics (e.g., interoperability and VVUQ), formulating working groups and
project teams to develop the new parts of the Digital Twin standards, and enhancing
the Digital Twin Lab to serve as a digital twin testbed to support digital twin
prototyping and standards development and testing.
Digital Twins for Advanced Manufacturing: The Standardized Approach                           167
Disclaimer Certain commercial products and systems are identified in this chapter to facilitate
understanding. Such identification does not imply that these software systems are necessarily the
best available for the purpose. No approval or endorsement of any commercial product by NIST is
intended or implied.
References
 1. Allen D (2021) Digital twins and living models at NASA [Slides]. In: ASME digital twin
    summit. https://ntrs.nasa.gov/api/citations/20210023699/downloads/ASME%20Digital%20T
    win%20Summit%20Keynote_final.pdf
 2. Grieves M (2002) Completing the cycle: using PLM information in the sales and service
    functions [Slides]. In: SME management forum. Troy, MI
 3. Khan A, Shahid F, Maple C, Ahmad A, Jeon G (2022) Toward smart manufacturing using
    spiral digital twin framework and twin chain. IEEE Trans Industr Inf 18(2):1359–1366
 4. Negri E, Berardi S, Fumagalli L, Macchi M (2020) MES-integrated digital twin frameworks.
    J Manuf Syst 56:58–71
 5. Jain S, Narayanan A (2023) Digital twin–enabled machine learning for smart manufacturing.
    Smart Sustain Manuf Syst 7(1):111–128
 6. Galli E, Fani V, Bandinelli R, Lacroix S, Le Duigou J, Eynard B, Godart X (2023) Literature
    review and comparison of Digital Twin frameworks in manufacturing. Proc Eur Counc Model
    Simul 2023:428–434
 7. Grieves M, Vickers J (2017) Digital Twin: Mitigating unpredictable, undesirable emergent
    behavior in complex systems. In: Kahlen F-J, Flumerfelt S, Alves A (A c. Di) Transdisciplinary
    perspectives on complex systems. Springer International Publishing, pp 85–113. https://doi.
    org/10.1007/978-3-319-38756-7_4
 8. Lu Y, Liu C, Kevin I, Wang K, Huang H, Xu X (2020) Digital Twin-driven smart manufacturing:
    connotation, reference model, applications and research issues. Robot Comput-Integr Manuf
    61:101837
 9. Accenture (2021) Think thread first: surf the wave of product data. Retrieved 6 Oct 2023, from
    https://www.accenture.com/us-en/insights/industry-x/thread-first-thinking
10. Singh V, Willcox KE (2018) Engineering design with digital thread. AIAA J 56(11):4515–4528
11. Shao G, Jain S, Shin J (2014) Data analytics using simulation for smart manufacturing. In:
    Tolk A, Diallo SY, Ryzhov IO, Yilmaz L, Buckley S, Miller JA (eds) Proceeding of the 2014
    winter simulation conference. IEEE, pp 2192–2203
12. ISO-23247-1 (2021) Automation systems and integration—Digital Twin framework for manu-
    facturing—Part 1: overview and general principles. Retrieved on 10 Oct 2023, from https://
    www.iso.org/standard/75066.html
13. ISO-10303 (2021) Industrial automation systems and integration—product data representation
    and exchange—Part 1: Overview and fundamental principles, Retrieved on 21 Oct 2023, from
    https://www.iso.org/standard/72237.html
14. MTConnect (2022) MTConnect standardizes factory device data. https://www.mtconnect.org/
15. ISO (2020) ISO 23952-2020. Automation systems and integration—Quality information frame-
    work (QIF)—An integrated model for manufacturing quality information. Retrieved on 11 Oct
    2023, from https://www.iso.org/standard/77461.html
16. ISO-23247-2 (2021) Automation systems and integration—Digital twin framework for manu-
    facturing —Part 2: reference architecture. Retrieved on 10 Oct 2023, from https://www.iso.
    org/obp/ui/en/#iso:std:iso:23247:-2:ed-1:v1:en
17. STEP Tools (2020) Demonstration of Three ISO 23247. Retrieved on 30 Nov 2023 from https://
    www.youtube.com/watch?v=wbsC_qzB8us
168                                                                                  G. Shao et al.
18. IEC-62832 (2020) Industrial process measurement, control and automation—digital factory
    framework—Part1: General principles, Retrieved on 12 Oct 2023, from https://webstore.iec.
    ch/publication/65858
19. IEEE-P2806 (2019) System architecture of digital representation for physical objects in factory
    environments, Retrieved on 15 Oct 2023, from https://standards.ieee.org/project/2806.html
20. IEC 63278 -1. (2023) AAS (Asset Administration Shell) for industrial applications - Part 1:
    Asset Administration Shell structure. https://webstore.iec.ch/en/publication/65628
21. ISO/IEC-30141 (2018) Internet of Things (IoT)—reference architecture, Retrieved on 11 Oct
    2023, from https://www.iso.org/standard/65695.html
22. ISO/IEC-21823-1 (2019) Internet of things (IoT)—interoperability for IoT systems—Part 1:
    framework, Retrieved on 11 Oct 2023, from https://www.iso.org/standard/71885.html
23. ISO/IEC/IEEE-15288 (2015)Systems and software engineering—system life cycle processes,
    Retrieved on 24 Oct 2023, from https://www.iso.org/standard/63711.html
24. DTDL(Digital Twin Definition Language) (2022) Digital twin definition language (DTDL)
    for models. Retrieved on 29 Sept 2023, from https://learn.microsoft.com/en-us/azure/digital-
    twins/concepts-models#digital-twin-definition-language-dtdl-for-models
25. HLA (High Level Architecture) (2010) IEEE standard for modeling and simulation (M&S)
    High level architecture (HLA)—federate interface specification. IEEE Comput Soc
26. OPC-UA (2017) OPC-UA services specification, Retrieved on 11 Oct 2023 from https://
    opcfoundation.org/developer-tools/specifications-unified-architecture#:~:text=OPC%20Unif
    ied%20Architecture%20Specification,more%20secure%20and%20scalable%20solution
27. MTConnect-OPC UA (2019) OPC-UA companion specification for MT connect. https://www.
    mtconnect.org/opc-ua-companion-specification
28. ISO/IEC-20922 (2016) Information technology—Message queuing telemetry transport
    (MQTT) v3.1.1, Retrieved on 26 Oct 2023, from https://www.iso.org/standard/69466.html
29. ISO/IEC-17826 (2016) Information technology—cloud data management interface (CDMI).
    Retrieved on 26 Oct 2023, from https://www.iso.org/standard/70226.html
30. ISO-13374 (2007) Condition monitoring and diagnostics of machines—Data processing,
    communication and presentation—Part 2: Data processing, Retrieved on 21 Oct 2023. From
    https://www.iso.org/standard/36645.html
31. ISO/IEC-30161 (2022) Internet of Things (IoT)—requirements of IoT data exchange platform
    for various IoT services, Retrieved on 23 Oct 2023, from https://www.iso.org/standard/53281.
    html
32. Automation ML (2021) Standard data exchange in the engineering process of production
    systems, Retrieved on 10 Oct 2023, from https://www.automationml.org/wp-content/uploads/
    2021/06/AutomationML-Brochure.pdf
33. CMSD (2010) Core manufacturing simulation data (CMSD) standard, https://www.sisostds.
    org/DesktopModules/Bring2mind/DMX/API/Entries/Download?Command=Core_Down
    load&EntryId=31457&PortalId=0&TabId=105
34. ISO-15531 (2017) Industrial automation systems and integration—industrial manufacturing
    management data—Part 44: Information modelling for shop floor data acquisition, Retrieved
    on 26 Oct 2023, from https://www.iso.org/standard/71064.html
35. ISO-14649 (2011) Industrial automation systems and integration—physical device control—
    data model for computerized numerical controllers—Part 201: Machine tool data for cutting
    processes, Retrieved on 26 Oct 2023, from https://www.iso.org/standard/60042.html
36. ISO/IEC-21823-2 (2020) Internet of things (IoT)—interoperability for IoT systems—Part
    2: Transport interoperability, Retrieved on 11 Oct 2023, from https://www.iso.org/standard/
    80986.html
37. ISO/IEC-21823-3 (2021) Internet of things (IoT) - interoperability for IoT systems—Part 3:
    semantic interoperability, https://www.iso.org/standard/83752.html
38. ISO/IEC-21823-4 (2022) Internet of things (IoT)—interoperability for IoT systems—Part 4:
    Syntactic interoperability, Retrieved on 11 Oct 2023, from https://www.iso.org/standard/84773.
    html
Digital Twins for Advanced Manufacturing: The Standardized Approach                              169
39. ASME (American Society of Mechanical Engineers) (1989) Digital representation for commu-
    nication of product definition data. Retrieved on 10 Oct 2023, from https://standards.globalspec.
    com/std/437642/ASME%20Y14.26M
40. ISO-10303-105 (2014) Industrial automation systems and integration—product data represen-
    tation and exchange—Part 105: Integrated application resource: Kinematics, Retrieved on 21
    Oct 2023, from https://www.iso.org/standard/64294.html
41. ASME (American Society of Mechanical Engineers) (2009) Data specification for properties
    of machining and turning centers, ASME B5.59-2
42. ISO-13399 (2006) Cutting tool data representation and exchange—Part 1: overview, funda-
    mental principles and general information model, Retrieved on 26 Oct 2023, from https://www.
    iso.org/standard/36757.html
43. ISO-16400 (2020) Automation systems and integration—equipment behavior catalogues for
    virtual production system—Part 1: overview, Retrieved on 26 Oct 2023, from https://www.iso.
    org/standard/73384.html
44. PMML (2018) The predictive model markup language (PMML) 4.3, Retrieved on 11 Oct 2023
    from http://dmg.org/pmml/v4-3/GeneralStructure.html
45. Zhang L, Zhou L, Horn BK (2021) Building a right Digital Twin with model engineering. J
    Manuf Syst 59:151–164
46. Hua EY, Lazarova-Molnar S, Francis DP (2022) Validation of digital twins: challenges and
    opportunities. In: Feng B, Pedrielli G, Peng Y, Shashaani S, Song E, Corlu C G, Lee L H, Chew
    E P, Roeder T, and Lendermann P (eds) Proceedings of the 2022 winter simulation conference.
    Singapore, pp 2900–2911
47. Kibira D, Weiss BA (2022) Towards a digital twin of a robot workcell to support prognostics
    and health management. In: Feng B, Pedrielli G, Peng Y, Shashaani S, Song E, Corlu CG,
    Lee LH, Chew EP, Roeder T, Lendermann P (eds) Proceedings of the 2022 winter simulation
    conference. IEEE, pp 2968–2979
48. ASME V&V 10-2019 (2023) Standard for verification and validation in computational Solid
    Mechanics. Retrieved on 11 Oct 2023, from https://webstore.ansi.org/standards/asme/asme10
    2019
49. ASME V&V 20-2009 (2023) Standard for verification and validation in computational Fluid
    Dynamics and Heat Transfer. Retrieved on 11 Oct 2023, from https://www.asme.org/codes-
    standards/find-codes-standards/v-v-20-standard-verification-validation-computational-fluid-
    dynamics-heat-transfer
50. ASME V&V 40-2018 (2023) Assessing credibility of computational modeling through
    verification and validation: Application to medical devices. Retrieved on 11 Oct
    2023, from https://www.asme.org/codes-standards/find-codes-standards/v-v-40-assessing-cre
    dibility-computational-modeling-verification-validation-application-medical-devices
51. Shao G, Frechette S, Srinivasan V (2023) An analysis of the new ISO 23247 series of stan-
    dards on digital twin framework for manufacturing. In: Proceedings of the 2023 international
    manufacturing science and engineering conference
52. Hussain M, Masoudi N, Mocko G, Paredis C (2022) Approaches for simulation model reuse
    in systems design—A review. SAE Int J Adv Curr Practices Mobility 4(2022-01-0355), pp
    1457–1471
53. National Academies, Sciences, Engineering, Medicine (2023) Foundational research gaps and
    future directions for digital twins. National Academies Press, Washington, DC. http://nap.nat
    ionalacademies.org/26894
54. Grieves M (2006) Product lifecycle management: Driving the next generation of lean thinking.
    McGraw Hill, New York
55. Pérez L, Rodríguez-Jiménez S, Rodríguez N, Usamentiaga R, García DF (2020) Digital Twin
    and virtual reality-based methodology for multi-robot manufacturing cell commissioning. Appl
    Sci 10:3633
NASA’s Michoud Assembly Facility:
Developing a Factory Digital Twin
Abstract This chapter presents a brief introduction and history of NASA’s Michoud
Assembly Facility (MAF) and current ongoing initiatives to reduce cost, improve
efficiency, and eliminate deficiencies/rework using Digital Twin (DT) technologies.
NASA’s mission to develop technologies and vehicles to literally reach for the stars
is by its nature very costly when looked at on a per-unit basis. Commercial space
flight companies, both in competition with and in partnership with NASA, are chal-
lenging NASA to explore ways to continue its unique mission while driving down
costs, particularly as these complex programs like the Artemis rocket and its compo-
nents transition from development to production. DT technologies have proven to be
effective in many commercial applications, as well as programs for the Department
of Defense, and hold substantial potential to help NASA achieve needed production
enhancements and cost reductions. The Michoud Assembly Facility is serving as a
prototype Factory DT for NASA as described in this chapter.
G. Porter (B)
Sev1Tech, Woodbridge, USA
e-mail: [email protected]
J. Vickers
NASA, Huntsville, USA
e-mail: [email protected]
R. Savoie
Loyola University New Orleans, New Orleans, USA
e-mail: [email protected]
M. Aubanel
Louisiana State University, Baton Rouge, USA
e-mail: [email protected]
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2024          171
M. Grieves and E. Hua (eds.), Digital Twins, Simulation, and the Metaverse, Simulation
Foundations, Methods and Applications, https://doi.org/10.1007/978-3-031-69107-2_8
172                                                                      G. Porter et al.
1 Introduction
Throughout its storied history, NASA has faced three major challenges that most
organizations do not have. These challenges are: (a) the systems they create are very
expensive; (b) they make very few of these systems; and (c) the systems they make
have not been made before [1]. The difference confronting NASA, and particularly
NASA’s human space flight programs today, is the recent development of a commer-
cial space industry and the successes achieved by some of the companies in that
industry.
    NASA is no longer competing only with other nations like Russia and China.
NASA now finds itself both competing with and partnering with commercial space
flight companies, particularly in Low Earth Orbit (LEO). SpaceX, for example, now
makes regular flights to the International Space Station, which is located in LEO
approximately 240 miles (400 km) above the earth. As such, it is no longer necessary
for NASA to build a separate spacecraft capacity to service LEO and compete with
these successful commercial space companies. In this case, the three issues delineated
above are no longer applicable.
    NASA is now focused on the next steps where the three challenges are still appli-
cable—returning to the moon to establish a permanent lunar presence and eventually
sending the first humans to mars. The moon is approximately 240,000 miles from the
earth, which is a totally different challenge than traversing the 240 miles to LEO and
mars is 230 million miles from earth, which is a massive step even from the moon.
    These are the type of challenges that fall into NASA’s mission space. The devel-
opment of the Space Launch System (SLS), the Orion Capsule and the Exploration
Upper Stage (EUS), all components of the Artemis rocket that have already circled
the moon, fit within NASA’s three key challenges. However, as SLS, Orion, EUS,
and other elements of the Artemis program move from development to production,
and as commercial space companies continue to develop more and more capabili-
ties, often in partnership with NASA, there is growing pressure on NASA to bring
down the costs of these complex systems to maintain its outreach to the moon, mars
and beyond. The DT technologies described in this chapter are one set of effectual
technologies that hold substantial promise for helping NASA to enhance production
and reduce the costs associated with continuing its mission to reach for the stars.
NASA’s Michoud Assembly Facility: Developing a Factory Digital Twin                           173
The NASA Michoud Assembly Facility (MAF) in New Orleans, Louisiana, has a
long and storied history. It was originally built in 1940 by the Higgins Industries
company as a production facility for cargo aircraft and tank engines during World
War II. NASA chose MAF in 1961 as the site for the manufacturing and assembly
of the first stages of the Saturn I and Saturn V launch vehicles (Fig. 1), which were
used to send astronauts to the moon during the Apollo program. MAF also played a
key role in the construction of the Space Shuttle, producing the external tank (Fig. 2)
from the first launch in 1981 to the mission in 2011. NASA’s space shuttle fleet flew
135 missions and was the delivery vehicle to support construction of the International
Space Station.
    In recent years, MAF has been involved in the development of new space explo-
ration technologies, including the Orion spacecraft (Fig. 3) and the Space Launch
System (SLS) (Fig. 4). MAF is one of the largest manufacturing facilities in the
world, with over 800 acres of land and 40 acres of manufacturing space under one
roof. Today, MAF is a vital asset to NASA’s space exploration program and to the
state of Louisiana. It is a symbol of American ingenuity and innovation, and it will
continue to play a key role in the future of space exploration.
Fig. 1 First stages of Saturn V rockets being assembled at the Michoud factory in the 1960s
174                                                                         G. Porter et al.
Fig. 3 Readying Orion for Flight, The NASA team at the Michoud Assembly Facility in New
Orleans has completed the final weld on the first space-bound Orion Capsule
NASA’s Michoud Assembly Facility: Developing a Factory Digital Twin                        175
Fig. 4 NASA Attaches First of 4 RS-25 Engines to Artemis I Rocket Stage Engineers and tech-
nicians at NASA’s Michoud Assembly Facility in New Orleans have structurally mated the first of
four RS-25 engines to the core stage for NASA’s Space Launch System rocket that will help power
the first Artemis mission to the moon
    New programs have allowed MAF to evolve and adopt new physical infrastructure
and technologies where needed and in an incremental fashion to meet program needs
such as the largest friction stir welding tool in the world, built for SLS and located
in the MAF Vertical Assembly Center (Fig. 5). However, legacy facilities like MAF
pose significant challenges to the adoption of widespread new technologies, hindering
innovation and progress. Government and industry alike have failed to focus on the
challenge of how to bring innovation to legacy sectors [2].
    These legacy sectors make up the majority of the US economy. There are nearly
300,000 factories in the USA, of which 90% have fewer than 100 employees.
Research indicates that it’s likely less than 10% of small and medium-sized manu-
facturing enterprises in the USA are extensively adopting emerging technologies
[3].
    Legacy facilities are often designed to accommodate outdated technologies and
processes, making it difficult to integrate new equipment or infrastructure. Physical
limitations such as limited space, outdated electrical and computer systems, or inad-
equate support infrastructure, can render the implementation of new technologies
impractical or extremely difficult.
    Cultural inertia is often one of the biggest challenges. Established practices and
routines within legacy facilities can create a culture resistant to change. Employees
accustomed to traditional methods may be hesitant to embrace new technologies,
even if they offer potential benefits. To overcome these challenges, organizations
176                                                                                  G. Porter et al.
Fig. 5 170-foot-tall, 78-foot-wide world-class welding tool that will be used to build the core stage
of America’s next great rocket, the Space Launch System
NASA’s Michoud Assembly Facility: Developing a Factory Digital Twin                 177
must adopt a strategic approach to technology adoption and foster a culture of inno-
vation. To remain competitive and not become obsolete, organizations must invest in
new technology and processes and allocate sufficient resources to acquire, implement,
and integrate new technologies. Organizations can encourage employees to embrace
new ideas, provide opportunities for experimentation, reward successful innovation
efforts, and create a supportive environment. Finally, they might partner with consul-
tants, or research institutions to gain expertise, access cutting-edge solutions, and
navigate implementation challenges.
    Today, these legacy challenges cause space systems development to take too long
and cost too much, and this high cost of operating in the space industry limits new
entrants and competitiveness. NASA and industry must deliver ever-more capable
products at reasonable cost. Manufacturing technology advancements must play a
critical role to improve US industry leadership in today’s hypercompetitive global
environment. The future of space manufacturing will be shaped by the digital trans-
formation and particularly by a new concept of radical innovation in design and
manufacturing called the DT.
    While the terminology has changed over time, the basic concept of the DT model
has remained fairly stable from its inception in 2002 [1]. The concept of the DT
dates back to a presentation at an SME conference in 2002 by Dr. Michael Grieves
[4]. The term DT was later coined by John Vickers around the year 2010 [5], and
because of their working relationship was attached to Dr. Grieves concept.
    The advancement of the DT approach has been tremendous. Today a Google
Scholar search yields more than 80,000 results and a Google search produces more
than 25,000,000 results. The DT approach has also evolved from its origins in manu-
facturing to vast applications in defense, business, transportation, science, medicine,
and a DT of the earth’s climate which is maybe the most complex system in existence.
More than anyone else, Dr. Grieves continues to shape the landscape of DTs.
    To propel MAF into the future NASA has awarded Louisiana State University
(LSU) a grant to produce the first real-time virtual representation, a DT of MAF. The
DT of the facility will mirror its real-life counterpart and replace the physical trial-
and-error approaches with computational data-driven modeling to improve design,
reduce cost, and improve quality.
    The MAF DT will also replace the outdated model room shown in Fig. 6.
Scale modeling or Replica Twins [6] has existed for millennia. Leonardo Da Vinci
was famous for creating intricate scale models of catapults, paddleboats, and even
mechanized robots during his life to present to the local rulers [7].
    Today, the scale model for utilitarian purposes is likely obsolete. LSU will utilize
the same technology used by the entertainment industry. The LSU College of Art &
Design and the LSU Digital Media Arts & Engineering program will lead the
construction of the DT that will eventually be housed at the LSU Campus. This
involvement of the Art department comes from a lesson from the Disney Imagineers
that don’t allow the engineers to participate in the creation phase of a major attrac-
tion until the artists have produced the concept. The reason is that engineers would
compulsorily favor physics over imagination.
178                                                                       G. Porter et al.
   The MAF DT will be an incredibly valuable tool, both for space industry profes-
sionals, who will be able to test a design at our facility from anywhere in the world,
as well as for educators, who will be able to bring the excitement of spacecraft
engineering to their students.
The benefits that may be realized from DT, digital thread, and Augmented Reality/
Virtual Reality (AR/VR) technology depend on the use cases specific to particular
factory applications, equipment, manufacturing processes/environment, challenges,
and desired outcomes. For MAF, and particularly the production of the SLS and EUS,
the initial challenges were due to the sheer size of the MAF manufacturing facility,
primarily building 103, and the massive size of the SLS and EUS “finished products.”
The initial steps were to combine laser scanning of the facility, equipment, and tooling
with the design of the SLS and EUS to ensure the “product” could be manufactured
at MAF without impediments such as transferring parts or a completed tank through
the various openings/doorways in the factory without hitting any obstacles that could
halt production. This step has been underway for years and provided an excellent
starting point for the work described later. While this work did not constitute a true
DT, it certainly made clear the benefits and cost savings that could be achieved by
generating a DT of MAF and the technologies associated therewith. The items listed
NASA’s Michoud Assembly Facility: Developing a Factory Digital Twin                179
below provide a high-level description of the potential benefits that can be realized
at MAF.
. Environmental Interrogation—floor levelness, floor load, environmental sensors,
  floor shifting with the tide. The floor of MAF is concrete and covers 40 acres.
  Since it was built basically at sea level near the coastal marshes of Louisiana, the
  floor moves up and down slightly with the tide. Further, the floor is not uniform or
  completely level. Understanding the movement of the floor and scanning the floor
  to identify low spots where moisture or water might gather is invaluable when
  planning the layout of the production process and understanding the environmental
  effects on equipment and components being manufactured.
. Simulation—cranes, manufacturing processes—cranes are but one type of basic
  tool that is needed in the manufacturing process. The location, lift capacity, main-
  tenance status, etc., are all critical to the process. Cranes must also be considered
  when planning the movement of large parts since they tend to be lower than the
  ceiling height and could pose an impediment.
. Factory optimization—tracking parts movement over time (discussed in detail
  later)
. AR for remote locations such as clean rooms—having a technician or other
  personnel use AR when performing work in a hard-to-reach location, such as
  a clean room, allows engineers and quality personnel to see what the technician
  sees without taking the time to enter the clean room or reach the remote location.
. AR/VR-based training and work instructions—Reducing non-conformances and
  rework by as much as 50%, as has been proven in other aerospace-related
  production facilities [8].
. Predictive/Preventative maintenance—For cranes, tooling, welders, etc.
. NASA Metaverse—Connecting all NASA sites, monitoring in real-time rockets
  or rocket sections being shipped from MAF to Kennedy or between any other
  NASA Centers. Supply Chain monitoring across all NASA sites.
4 Gamification of DTs
One of the largest shifts in how we interact with digital artifacts comes down to real-
time interaction with real-world physics and data in fully realistic 3D environments.
Automobile design, simulating crash tests, architecture, medical visualization, manu-
facturing processes, and training are all being revolutionized by real-time engines.
This allows engineers and designers to manipulate and design via computers in real
time.
   Two of the popular video gaming development engines are Unreal and Unity.
However, both game engines are used extensively outside of the video game devel-
opment field where these engines were initially conceived. In 2022, the total revenue
of the video game industry in the USA reached an impressive $56.6 Billion [9]. As
per a report from the Entertainment Software Association unveiled on July 10th,
2023, 65% of the American population engages in playing video games, with a
180                                                                        G. Porter et al.
represent the entire facility. These challenges led us to select Unreal 5 game engine
as the backend platform for the current iteration of the DT platform. There are a
few features in Unreal 5 that we found beneficial in overcoming these challenges:
Nanites and Datasmith imports.
    Nanites replaces regular geometry within the gaming engine. Rendering using
Nanites is a fundamental shift in how to make a level, or in our case a DT, performant.
Traditionally, for 3D simulations we are “draw call bound” which results in a finite
number of polygons that the GPU can keep in memory. The ability or inability to
render a level in a game engine is the direct result of how many polygons are present.
In large environments such as the recreation of MAF, software developers would have
to use many tricks and complex techniques to lower the polygon count. However,
since Nanites is screen resolution bound and not “draw call bound,” we can have a
much larger polygonal count.
    For example, if we are 1000 feet away from a piece of equipment within the DT,
the model may only consist of 10,000 polygons. But, as we approach the equipment,
the polygons increase significantly to provide an extremely realistic representation of
the equipment. A large piece of equipment can easily contain millions of polygons.
Ultimately, this allows us to achieve high accuracy and model at the highest level
of detail (LOD) possible. The engine will scale the model automatically to what is
needed to represent it realistically on screen. This is a game changer for us as it
allows us to generate a single model. With prior engines, we had to represent objects
with multiple levels of detail, as the engine now takes care of the LOD.
    Based on previous experiences, we knew one of the large challenges we would
face on the project was working with engineers and architects who use CAD data
to be able to view their work in a polygonal real-time environment. However, the
compatibility between CAD and polygonal models is poorly supported. Additionally,
the techniques to design a piece of manufacturing equipment and a game differ
greatly. When designing manufacturing equipment, it is normal to create the entire
model in great detail, as a single model. In game development, modeling is typically
broken into small pieces and then placed together, analogous to building using Lego
bricks to build a model versus having a single piece be the whole model.
    The solution to this problem is Datasmith. It allows us to take models created
in CAD and split them into smaller polygonal objects that we can then directly
import into Unreal 5. This closes the gap between the DT platform’s development
workflow and the CAD modeler’s workflow by increasing collaboration and reducing
the amount of rework needed. Furthermore, by incorporating these tools and data
objects into the digital thread (discussed in Sect. 6), we can fully understand and
track the relational dependencies.
    By using Unreal Engine 5, we also help solve another critical problem: cyber
security. By making the DT platform widely available to many different devices such
as computers, cell phones, tablets, and AR/VR gear, and people, there is an increased
risk of a data breach. To alleviate these cyber security concerns, we are using pixel
streaming. Pixel streaming allows us to run the DT platform from remote servers
which prevents users from having to download the software. Instead, operators just
182                                                                         G. Porter et al.
connect to the DT platform with a web browser (this topic is discussed in greater
detail in Sect. 7).
    Another key factor for the project is to create a user interface that is easy to use,
while allowing immersive interactions. To do this, the team is relying on sandbox-
styled games like SimCity (Maxis). The goal is to create an extremely detailed
environment with many base features to drive engagement by the operators.
    Our reliance predominantly rests on sandbox-style games such as SimCity (Maxis)
for our DT development. Our emphasis lies in fostering experimentation and facili-
tating changes within the operational facility. Despite representing intricate details,
our goal is to encourage operators to experiment and craft scenarios that go beyond
the scope envisioned by the DT’s software designers. This approach not only ensures
simplicity in controlling a complex system but also poses a primary challenge for
our team: crafting a user interface that strikes a balance between ease of use and
enabling robust interaction.
    Through the creation of tools designed to address envisioned use cases, we antic-
ipate diverse applications beyond the team’s initial considerations. Drawing inspi-
ration from the transformation witnessed in the video game industry, where devel-
opers released products before completion, allowing players to contribute to the final
version of the game, our approach is similarly dynamic. The DT serves as a catalyst for
this project team to significantly alter NASA’s operational landscape, contributing
to the overarching objectives of achieving lower costs, enhanced reliability, and
ultimately, the development of an improved rocket for future mars missions.
In the pursuit of precision, imperative when building rockets, our 3D laser scanning
techniques to capture the “as-built” state of the factory, equipment, tooling, and
flight hardware are highly refined and continually tested. The utilization of laser
scanning provides a meticulous and highly accurate method for spatial data capture,
ensuring modelers have a robust foundation to work with. The incorporation of
target-based registration, which is the practice of using artificial targets, or points
and planes extracted from scan points, to align multiple scans, further refines this
process, increasing the alignment and accuracy of the captured data. This precise
dataset serves as a crucial starting point for our 3D modelers, providing them with a
detailed and reliable representation of the physical environment through the resulting
point clouds.
    In looking for ways to increase efficiency, our DT platform introduces a unique
capability: the ability to display point clouds without the necessity of 3D models,
also referred to as meshes. This feature proves invaluable in scenarios where time
and cost considerations dictate the omission of a full 3D model. Point clouds within
the DT retain all the functionality and richness of a traditional 3D model, providing
a comprehensive and immersive experience. However, it is crucial to acknowledge
a tradeoff: while point clouds offer a time-saving alternative for visualizing the “as-
built” world, they require more computational power than their 3D model (mesh)
counterparts. The increased computational requirements underscore the importance
of balancing modeling intricacy with available compute resources.
    Advancement in technology has paved the way for increased utilization and explo-
ration of photogrammetry, defined as the process of deriving metric information
about an object through measurements made on photographs of the object [13].
With high-definition photographs serving as the data source, it is applicable, at low
costs, to capture 3D morphology, including photo-realistic textures in high reso-
lution with large overlaps. Our modelers use this during modeling to implement
realistic textures and color palettes to create Level of Detail (LOD) 400–500 meshes
and textures. Ultimately, the goal of photogrammetry is automatic generation of 3D
models from photographs, resulting in a significant leap in reality capture capabili-
ties. However, current practices often require manual interactions such as geometric
modeling, object identification, and monitoring [14]. Despite the rapid strides in this
exciting subset of reality capture, which may one day replace most 3D modeling for
as-built objects and environments, we believe that, for the precision required in most
areas of our project, photogrammetry has not yet reached its full potential. We will
keep experimenting with and exploring the technology.
184                                                                        G. Porter et al.
The transition from point clouds to 3D models is where our efficiency begins to take
center stage. Using Machine Learning (ML) models, discussed in detail within the
“Artificial Intelligence, Machine Learning, and DTs” section, we can automatically
extract objects from point clouds. This process greatly reduces the time 3D modelers
must spend trying to manually separate objects from much larger point clouds. The
smaller object-oriented point cloud segmentations allow the 3D modelers to produce
highly accurate 3D models that accurately replicate the physical environment. The
conversion process of smaller, object-oriented point cloud segmentations into 3D
models is not only swift, but remarkably precise, enabling the creation of virtual
replicas that mirror reality in astonishing detail.
   These 3D models within the DT platform serve as a virtual representation of the
physical environment, offering a wealth of insights for decision-making processes
and coordination; NASA and its contractors can leverage to conduct spatial analysis,
simulate various scenarios, and optimize designs. Additionally, the intricate details
captured through laser scanning later translated into 3D models empower decision-
makers with a holistic understanding of the environment, allowing for informed data-
driven choices in various aspects of project planning, execution, and maintenance.
the DT. By precisely tracking the location and movement of assets, the DT can
simulate and predict scenarios with a high degree of accuracy, offering a powerful
resource for scenario planning, predictive maintenance, complex crane lifts, and
operational optimization.
   The relational linking of sensor tags to other data systems through the digital
thread is a critical aspect of this integration. It creates a flow of information allowing
the DT to tap into disparate systems containing details about the assets being tracked,
prior/future work orders, active capacities, and much more.
MIT defines a digital thread as a data-driven architecture that links together infor-
mation generated from all stages of the product lifecycle and is envisioned to be the
primary or authoritative data and communication platform for a company’s products
at any instance of time [15].
    At the practical level, implementing the digital thread augments existing enterprise
integrations. This evolution is based on advancements within information technology,
driven by well-known and used innovations such as cost-effective and nearly limitless
computation and storage capabilities through cloud computing. Additionally, lighter,
and more reliable data transmissions, inspired by optimizations realized in internet
streaming, contribute to the seamless flow of information between systems. This
contemporary landscape is further shaped by faster and stronger databases, new ways
of controlling versions from disparate systems, affordable and dependable sensors,
and the application of sophisticated techniques in physical object scanning and image
processing.
    The integration of these innovations not only strengthens the existing data infras-
tructure but also enriches the product development process measurably. The digital
thread allows for comparison, including visualizations, of product data across diverse
lifecycle stages. For example, it facilitates the comparative analysis of “as-built”
digitalized tolerance measurements against the original design in a computer-aided
design (CAD) file. This integration introduces closed feedback loops, contributing to
a more comprehensive understanding and optimization of the product development
lifecycle.
    Value also arises from the automated provisioning of updated authoritative data
sources, making the digital thread itself the authoritative source. It allows data to
stay within its source system, exposing only necessary metadata until the actual
data is needed. This empowers subject matter experts to retain control over their
information. Unlike traditional data workflows requiring resource-intensive Extract,
Transform, Load (ETL) workflows to replicate data into a centralized data warehouse,
the digital thread renders these processes obsolete by allowing data to reside within
its originating system. This streamlined approach conserves organizational resources
and enhances agility and responsiveness.
NASA’s Michoud Assembly Facility: Developing a Factory Digital Twin                 187
   Beyond data control, digital threads have established themselves as the corner-
stone in DT architecture by facilitating the near real-time synchronization of digital
system models and their corresponding DTs. Establishing connections between
disparate systems fosters a relational understanding of the data, systematically
breaking down data silos and providing organizations with a comprehensive view
that leads to novel and insightful observations. The relationships between systems
can be created through ontologies, which represent a shared, explicit specification of
a conceptualization of the domain of knowledge [16]. It provides a formal specifi-
cation of the vocabulary of concepts and the relationships among them, in a domain
of interest [17].
   In the diverse ecosystem of NASA, comprising vendors, contractors, and tenants,
each entity operates with distinct software, processes, and scopes of work. To deploy
an effective digital thread in such an ecosystem, it must be platform agnostic, allowing
entities to continue their operations as usual. Complex systems can be interconnected
through innovative tools such as the digital thread, utilizing modern-day Application
Programming Interfaces (APIs). These APIs enable seamless interactions with the
digital thread, facilitating native integration and data exchange.
7 DT Accessibility
Cloud computing has become a linchpin for businesses seeking flexibility, cost-
efficiency, and scalability in their IT infrastructure. One pivotal advancement in cloud
computing is the availability of GPU resources, significantly enhancing the perfor-
mance of applications that require intensive graphical processing. This is particularly
beneficial for workloads such as ML, simulations, and realistic rendering of DTs. By
leveraging cloud services with GPU support, organizations can offload the burden
of managing the procurement and deployment of specialized hardware, ensuring
that resource-intensive tasks are executed efficiently. By leveraging cloud services,
188                                                                      G. Porter et al.
our project has been able to offload the burden of managing physical hardware,
allowing us to focus on developing automated and highly scalable DT platforms,
interconnected digital threads, and near real-time 3D model data catalogs.
   The pay-as-you-go model of cloud computing ensures cost-effectiveness, since
we only have to pay for the resources we consume. Moreover, the scalability of
cloud resources enables us to effortlessly handle fluctuations in demand by using
autoscaling and rightsizing best practices for modern-day architectures, ensuring
optimal performance during peak times, and cost savings during periods of reduced
activity. Cloud computing services have enabled us to innovate rapidly and remain
agile in today’s dynamic digital landscape. Furthermore, by using security groups,
we implicitly deny all access and explicitly allow access, giving us granular control
over infrastructure and applications.
The convergence of Kubernetes, cloud computing, and Pixel Streaming lays a robust
foundation for DTs. In this advancing technological landscape, the integration of
these tools is proving to drive innovation across not only NASA and aerospace indus-
tries alike, but also education, manufacturing, energy, defense, real estate, marketing,
and many others.
    By containerizing Pixel Streaming and deploying it on Kubernetes within a cloud
environment, we provide a consistent and portable DT that is buildable, testable,
and deployable through continuous integration/continuous deployment (CI/CD)
pipelines. Deploying on Kubernetes introduces unparalleled scalability, allowing us
to efficiently handle varying workloads and fluctuations in user demand, even scaling
down to zero resources when no one is using the DT platform. This dynamic scal-
ability, coupled with the inherent accessibility of cloud resources, ensures optimal
performance regardless of geographical location or personal hardware specifications.
    Ultimately, this leads to cost reduction by optimizing resource utilization and elim-
inating unnecessary expenses. The orchestrated collaboration between Kubernetes,
cloud computing, and Pixel Streaming reflects not only the current but also the future
state of technology, underscoring a vision where interactive, visually compelling
190                                                                       G. Porter et al.
In the dynamic landscape of DTs, the strategic integration of AI and ML not only
revolutionizes operational patterns but also serves as a jumping point for maxi-
mizing return on investment. As organizations increasingly harness the power of
DTs, the infusion of intelligent technologies amplifies their transformative potential.
This section delves into key applications within the MAF DT, focusing on our DT
assistant, point cloud segmentation, and predictive maintenance, all underpinned by
a central theme—the realization of substantial returns on investment. By exploring
how AI and ML elevate decision-making, streamline workflows, and enhance predic-
tive capabilities, we unveil the many ways in which NASA stands to gain significant
value and efficiency from their investments in these cutting-edge technologies.
Automated point cloud segmentation has emerged as a pivotal application within our
DT project, specifically in expediting the 3D modeling workflow. By using state-
of-the-art Learnable Region Growing for Class-Agnostic Point Cloud Segmentation
ML model, we have been able to automate the segmentation of point clouds. By
transforming extremely large, dense point cloud data into structured segmented point
clouds we have significantly reduced the time and effort traditionally associated with
manual segmentation.
The convergence of AI, ML, and DTs facilitates predictive maintenance strategies
that revolutionize Industrial Internet of Things (IIoT) applications. Capturing IIoT
data over time allows for a deep understanding of patterns within the data, enabling
the prediction of potential issues before they occur. This capability extends beyond
routine machinery maintenance, delving into complex scenarios such as predictive
crane maintenance and assessing the effects of environmental factors like temperature
and humidity on critical processes like welding and chemical applications.
   In summary, the harmonious integration of AI and ML technologies within DTs
transforms these virtual replicas into intelligent, dynamic entities. The DT Assistant,
point cloud segmentation, and predictive maintenance represent key applications,
showcasing the immense potential for enhanced decision-making, operational effi-
ciency, and innovation across diverse industries. This synergy underscores the pivotal
role of AI and ML in shaping the future trajectory of DT technologies.
9 Conclusion
In this chapter, we have covered the intricacies of the NASA Michoud Assembly
Facility (MAF) DT project, a groundbreaking initiative at the forefront of precision
engineering and exploration. As we conclude our exploration of the five key dimen-
sions: gamification of DTs, comprehensive eality capture of DTs, digital threads,
DT accessibility, AI, ML, and DTs of this comprehensive endeavor, it is evident
that the integration of cutting-edge technologies is reshaping the landscape of space
exploration and manufacturing.
   The historical context provides a foundation for understanding the evolution of the
Michoud facility and NASA’s rich legacy. This historical perspective underscores the
significance of adapting traditional facilities to the demands of contemporary space
programs, setting the stage for the transformative journey into the digital realm.
192                                                                              G. Porter et al.
References
Abstract This chapter overviews modeling and simulation methodology for smart
airport Digital Twin metaverse applications, emphasizing the development of a
distributed agent-based modeling implementation framework. It outlines current
approaches and applications of modeling and simulation concerning optimization,
sustainability, and digital transformation, along with theoretical, statistical, and
empirical approximations to model validation and accuracy assessment. It also pro-
vides a rudimentary Agent-Based Modeling (ABM) categorization and classifica-
tion of current models. We provide a key perspective to airport planning model
development, including resource allocation, workflow design principles, and process
optimization. An example reference and design of distributed and intelligent agent
architecture for a smart airport is provided, along with the design elements for agent
allocation, communication, and orchestration mechanisms for system architecture.
Finally, we provide a series of inferences, insights, and principles to guide the design,
K. Alexandridis (B)
Orange County Public Works, OC Survey Geospatial Services, Santa Ana, CA, USA
e-mail: [email protected]
S. Sabri
Urban Digital Twin Lab, School of Modeling, Simulation and Training,
University of Central Florida, Orlando, Florida, USA
e-mail: [email protected]
J. Smith
Sierra Nevada Corporation, Nashua, NH, USA
e-mail: [email protected]
B. Logan
Rockport Software, Slough, England, UK
e-mail: [email protected]
K. Bartfai-Walcott
Ambient Enterprises Inc., California, USA
e-mail: [email protected]
D. Migliori
Event Driven Systems, Irvine California, USA
e-mail: [email protected]
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2024          195
M. Grieves and E. Hua (eds.), Digital Twins, Simulation, and the Metaverse, Simulation
Foundations, Methods and Applications, https://doi.org/10.1007/978-3-031-69107-2_9
196                                                                  K. Alexandridis et al.
1 Introduction
A growing body of practical and scientific work recognizes the importance of Agent-
based Modeling (ABM) and simulation in enhancing transportation operations. For
example, passenger flow modeling is one of the widest applications of ABM in
terminal models, including railways and airports. This approach has been used for
many use cases, such as improving non-aeronautical revenue, passenger satisfaction,
hazard mitigation and evacuation, and energy efficiency [1, 2]. However, there are
limited scientific reports on system architectures that incorporate airport planning
and operations into a holistic digital environment. Furthermore, the current top-down
management approach in many airports and other transportation hubs overlooks the
human and social factors as essential dimensions in the operational planning and ser-
vice allocation [3]. Given a progressive increase in the deployment of sensors, mon-
itoring equipment, and high-performance computing technologies, the traditional
modeling, simulation, and prediction approaches could be enriched with real-time
and streamlined data.
    This chapter concerns optimized ABMs calibrated with real-time data. The cur-
rent state-of-the-art ABM supports a distributed AI-enabled metaverse, facilitating
transparency and flexibility of management and operations in complex environments
[4]. This method can also enable an adaptive system of systems (SoS) where stake-
holders collaborate and understand the complex implications of their decisions and
action scenarios.
    We propose a distributed AI-based modeling and simulation architecture for smart
airports. To do this, the chapter first explores the state-of-the-art applications of ABM
in airports. We will explore the current applications of modeling and simulation and
ABM in different dimensions, including human systems, physical and operational
aspects, and sustainability. We pay specific attention to the Modeling and Simulation
in transportation hubs in a subsection. Then, we investigate the validation approaches
such as spatial, temporal, systemic, and behavioral. These will be translated into our
proposed architectural principles. The next section describes an airport planning
model, the definition of its components, rules, resource criteria, and usage. Using
the content of this airport planning model, we propose a smart airport ABM design,
and then we provide the visual metaverse aspects of an intelligent Digital Twin
framework for smart Airports in the next section. Finally, the chapter provides the
concluding remarks and sets an agenda for the development and implementation of
a new generation of ABM integrated into the Digital Twin architecture to enhance
Airport Planning.
Distributed AI Modeling and Simulation for Smart Airport …                           197
Given the complexity of airport operations, this chapter focuses on three interre-
lated dimensions: (a) the human dimension (passenger flow); (b) the physical and
operational dimensions (terminals and airlines), and; (c) the sustainability dimension
(energy efficiency). In this section, we review and synthesize the current state of the
art in modeling, simulation, and virtual representation of airport operations, as well
as social and behavioral theories and validation methods to identify the progress and
limitations in the applied and scientific environments.
Several studies indicated the importance of physical and operational strategies for
passenger flow control. For example, increasing the service process capacity and
198                                                                   K. Alexandridis et al.
social distancing during the COVID outbreak informed a better intervention and
decision-making process in major transportation hubs [11]. Examining the inbound
passenger flow lines through the average inbound time of passengers from each
node (parking to the entrance and thereafter) and the average queue length for each
kind of passenger service equipment facilitates gaining an insight into the terminal
layouts [12]. The latter can also be modeled as the average service/processing time
of each service, including passenger service equipment, while average queue length
and wait times can be derived from combining service times and passenger volumes
interactively [13].
   Other studies looked into integrating pedestrian facilities planning and staff
assignment for transfer capacity, transfer average time, and level of service [14].
The passenger flow line regulated the entrance and transition gates, such as bidirec-
tional automatic fare gates (BAFGs), and avoided passenger congestion [15]. Given
the time-varying passenger demands, this operational problem could be improved
through real-time/near real-time data.
   Undoubtedly, the airlines, as main stakeholders, play a crucial role in airport
operations, and their operational model contributes to the smoothness of passenger
flow. The service process capacity and staff assignments mentioned above could
affect airlines’ operational strategies. Airline’s approach to passenger boarding has
implications for passengers’ satisfaction, the operational timing, and the energy used
to air condition the boarding bridges [16]. As such, the physical and operational
dimensions are interrelated to human activities and sustainability.
1.1.3 Sustainability
Airport terminals consume more energy than normal public buildings on average
[17]. Studies indicated that three factors are correlated to the airport terminal’s energy
consumption: (a) passenger flows, (b) meteorological parameters, and (c) supply fan
frequency (for the zonal airport terminal HVAC system) [18]. Given these factors,
the role of ABMs in understanding the environmentally sustainable performance of
airports is significant. We argue that the energy consumption of airport terminals
poses a substantial challenge to the carbon neutrality policy due to the high energy
demand and round-the-clock operation of these terminals. Integrating geospatial
data and analytics into the ABM allows the development of the spatio-temporal
passenger distribution model to describe the transient passenger distribution pattern
of an airport terminal [19, 20]. The outputs of such models are used as input into the
energy simulation tools to estimate the energy consumption of airport terminals in
different scenarios [21].
   Furthermore, passenger flow plays a vital role in the terminal’s energy consump-
tion and indoor environmental conditions [22]. A prediction model of passenger
flow, developed by [22], confirmed the operation optimization and terminal’s lay-
out design to address energy efficiency. Alternatively, energy consumption for each
service level can be aggregated based on overall service use [13].
Distributed AI Modeling and Simulation for Smart Airport …                          199
Recent research has highlighted the value of hybrid models in transportation hubs
and, specifically, airport studies [23]. To improve the reliability of models, the com-
bination of modeling and simulations and the integration of emerging technologies
such as Machine Learning (ML) and Artificial Intelligence (AI) for optimization and
predictive analysis are suggested.
   For example, the passenger movements, multi-gate and multi-destination nature,
passenger service processing nodes, and their operational patterns in airports justified
the hybrid ABM and discrete event system (DES) models to develop a multi-function
simulation tool to help inform decision/policymakers in different operational phases
of planning, design, development, and implementations for addressing issues like
public health emergencies and emergency evacuations [11].
   Example application integrated agent-based simulations and machine learning
methods, such as random forests, enable broader model optimization and estimation
robustness for transportation hub’s facility planning and staff assignment [24]. How-
ever, some challenges remained unresolved for the hybrid models due to the limited
availability of simulation platforms, high demand for computational resources, and
limitations in validation and optimization [11].
   In macro-to-micro simulation models as a group of hybrid models, agent micro-
behaviors and spatio-temporal movements are often combined to understand and
model broader system dynamics. For example, in an airport model [15], passenger
micro-behaviors and movements aid the understanding of system-wide efficiency
patterns and parameters while enabling optimized configuration or efficient arrange-
ment of system resources.
   In addition, models of flow dynamics in terms of spatial and temporal movements
and shifts are categorized as prediction models. Of particular importance is under-
standing optimization and flight arrangements to predict passenger flows in airport
terminals or using the reverse problem formulae, using predictive passenger flows
across space (terminal sections, gates, security points) and time (peak times, off-
peak times) to optimize resource and system provisions or estimate carrying/critical
capacity flows for the terminal system.
Often, model validation and accuracy issues fall within four related categories and
groups. Specifically, these can be (a) spatial or geostatistical, (b) temporal, (c) sys-
temic or engineering, and (d) behavioral or intelligence-related. Validation and reli-
ability represent a cornerstone for the evolution of metaverse in the future mobility
[25].
200                                                                 K. Alexandridis et al.
Spatial/Geostatistical
Normal statistical model validation and accuracy assessment concern what and to
what degree something occurs or emerges from the model behavior. Geostatistical
model validation approaches not only must incorporate standard statistical methods
but additionally must include where and to which scale such patterns and model
behaviors occur [26]. Issues of spatial accuracy often involve spatial statistics about
spatial autocorrelation, spatial heteroskedasticity, and scale variance (or invariance).
For example, validation patterns that pass accuracy testing at 1:100,000 scales might
fail at 1:10,000 scales and vice versa, rendering scaling up or down for model dynam-
ics problematic or completely fallacious. Modeling and simulation patterns of behav-
iors and results must often be able to stand scaling-up and down approaches and thus
require multi-resolution validation data or data sources that vary in spatial intensity
or at least sampled across a varying degree of spatial resolution.
Temporal
Systemic/Engineering
Behavioral/Intelligence
In many cases, micro-finite data patterns on individual behaviors are often desired and
needed to validate model estimations and results. For example, Scala et al. [28] used
historical data and data analytics methods to validate their modeling and simulation.
They generated a passenger “Activity-Travel-Diary” model from different sources,
such as passenger’s data (personal mobile phone, apps), airport data (airport WiFi,
GPS, scanning facilities), and flight information (flight schedules and gate alloca-
tion) and successfully validated the simulation results. Creative uses of existing data
and multi-source data pluralism can provide valuable insights into model behaviors
and predictive patterns emerging from modeling and simulation approaches. On the
other hand, behavioral and cognitive patterns and data cannot be used strictly to
establish causal predictive behaviors and require additional heuristic, probabilistic,
or propabilogic assessments and model validation methodologies. Agent-baMicro-
finite particularly sensitive in terms of statistical validation methods, as their design
and definitions encapsulate and embed a certain degree of dynamism, randomness,
and uncertainty. Heuristic approximations and assessment of realities also require
non-static methods of assessment and repeated measures such as statistical boot-
strapping, Monte Carlo simulations, or simulation ensemble methods.
This study focuses on two categories of agent modeling: (a) phenomena-based mod-
eling; and (b) exploratory modeling [29]. In phenomena-based modeling, one uses
agent rules to create known phenomena represented as a reference or aggregate pat-
tern. An example of this with respect to an airport planning model is to change a
traffic distribution pattern that may be derived from various sources including Airport
Operator or Air Traffic Management.
   In exploratory modeling, one creates a set of agents, using their behavior to explore
emergent patterns. An example of this exploratory modeling with respect to an airport
planning model is given in [30] where one simulates an airport with passenger agents
passing through several airport zones, varying parameters such as boarding gates,
shops, and arrival patterns, to observe the emerging behaviors.
   We expect to use both methods, with the exploratory modeling providing the basic
agent framework and parameters and the phenomena-based method used to replace
the simplified distribution patterns used in the exploratory model.
   As one refines a model with either modeling category, conformance to agent
modeling standards such as the Foundation for Intelligent Physical Agents1 (FIPA),
1   www.fipa.org.
202                                                                  K. Alexandridis et al.
1.2.2 Metaverse
1.2.3 Challenges
Most current approaches, however, have failed to provide a robust, reliable, and
scalable solution to incorporate real-time data in the calibration, validation, and sen-
sitivity of tools. The existing data collection methods often lead to biased data [41].
Most data used to calibrate the models are historical or experimental. Accordingly,
agent-based models such as passenger flow models are highly simplified. The lack
of data led to the introduction of many assumptions to the model, which decreased
their reliability. In addition, in many cases, the airport terminals are considered indi-
vidual entities, whereas these buildings are connected to broader and mass build-
ings/infrastructures and catchment areas in their precinct and city (e.g., parking, city
transportation).
2   fipa.org/repository/aclspecs.html.
Distributed AI Modeling and Simulation for Smart Airport …                                   203
Multi-agent architectures, in general, are ideal for linking Digital Twin use cases.
The multi-agent network on the “digital” side can be arranged to resemble the “real”
side structure. Each agent on the “digital” side can freely communicate with the
sensors on the “real” counterpart. Parallels between the Digital Twin and agents are
given in Table 1.
    In our case, the phenomena-based modeling part approximates actual airport traf-
fic with known patterns and varying initial parameters (e.g., in and out traffic arrival
rates, number of agents, boarding gates, shops, check-in boxes, baggage belts, and
station lag times). The exploratory-based modeling part tests the consequences of
using different interaction protocols, varying the same parameters. Many commer-
cial frameworks rapidly create agent applications using the Foundation for Intelli-
gent Physical Agents (FIPA) and Agent Communication Language (ACL) standards,
using tools, e.g., AgentSpeak, JACK, JADE, GAMA, Mason, Repast, and NetLogo.
Among these applications, NetLogo [42] provides an environment to rapidly under-
stand the value of ABM Digital Twin integration because it:
• Supports the FIPA and ACL standards through extensions;
• There is built-in support for connecting GIS data;
• Is being discussed in the Digital Twin Consortium, Mobility and Transportation
  Working Group, to implement an air traffic use case used to interact with different
  types of agents (e.g., positioner, evaluator, service provider, and user agents) and
  has been used for similar applications [30], and;
• Has human-like (Belief, Desire, and Intention) and communication extension code
  that could conform to the agent specification [43] by replacing existing NetLogo
  extensions.
This section provides the workflow of airport planning and operational models. These
workflows will create a basis for the smart airport Agent-based Modeling design and
implementation explained in Sect. 3.
Distributed AI Modeling and Simulation for Smart Airport …                           205
Many parties are involved in generating schedules for all activities across the airport.
The multiple interdependencies between schedules mean that it may take weeks or
months to develop a default schedule. The business rules that a schedule may need
to take into account are not always written down, making it more difficult and time-
consuming to generate the schedule and to verify it against operational scenarios.
    Airports operate within an ecosystem, with multiple parties providing and con-
suming services within the ecosystem. Whether a hub or point-to-point, an airport
acts as an intermodal transportation hub supporting a wide variety of transport modes,
which will increasingly consist of unmanned vehicles subject to automated control.
Interoperability in the SoS depends on the effective exchange of information based
on a common understanding (human and machine) between the parties, including
the airport, airlines, air traffic management, and a multitude of other participants in
the ecosystem. Common ontologies and associated data standards are fundamental
to meaningful information exchange.
    Consequently, airport planning is not conducted in isolation. The airport operator
provides information on demand and capacity to airlines. Control authorities such
as air traffic management, and local aviation authorities impose constraints such as
the minimum time between landing two aircraft of type F. Each organization has its
own capabilities, including landing aircraft, refueling aircraft, immigration, check-in,
baggage handling, airfield maintenance, etc. Other factors, such as climate conditions
(visibility level) and social events (see example in Table 2), will increase the level of
complexity in the operations.
    The capabilities to do work require resources such as aircraft and surface transport
vehicles, terminal facilities, airport equipment and facilities (runways, taxiways,
stands, etc.), and ground handlers. These resources not only need to be deployed but
they must also align with many other organizations. Planning activity occurs in many
different places in many different organizations and the interdependencies must link
for the operation to work effectively.
    All work in an airport operation occurs in an environment, usually consisting
of climatic conditions such as a working temperature range, and social conditions
such as train driver strike. Plans are created for an environmental scenario category.
Environment scenarios can change quickly.
206                                                                K. Alexandridis et al.
All organizations involved in an airport operation have their planning cycle typified
by the following diagram in Fig. 1. Four stages of the plan management cycle are
demand and capacity management, work scheduling, work execution, and perfor-
mance management. The ideal situation is a smooth flow of information across these
four stages.
    This high-level perspective helps develop planning activities’ processes and infor-
mation requirements. Considering the power of Digital Twin for real-time data
exchange and analytics and Metaverse for cross-collaboration in a virtual space, it
is crucial to understand the business process of airport operation, which is explained
here. Default schedules are created for all work activities based on a load-balanced
capacity profile in a specified environmental scenario, taking account of resource
usage constraints. The planner obtains the demand for airport resources from demand
and capacity management, including the scenario it operates within, for example, a
normal day. This constitutes the required capacity profile. Demand and capacity
management also provide the resource usage constraints for the scenario.
Distributed AI Modeling and Simulation for Smart Airport …                            207
   The planner defines the business rules for the airport, airline, and government
work units within the airport for the provided scenario, including the default work
dependencies, the default work criteria, and the default resources to execute the work
(for example, in Fig. 2).
   All schedules are verified against forecast assumptions, usually via a manual
process or automatically by process simulation or flow emulation. However, real-time
data is lacking in most of the cases. The outcome is a set of work units that do or do not
meet performance targets. Once default schedules have been verified and activated,
they typically need to be adjusted on the day to take account of circumstances such
as late arriving or departing flights, bad weather, and staff absence.
   Given the lack of real-time data analytics and a single point of truth for stakeholder
collaborations (Metaverse) changes to plans may take many hours, particularly if
they have to be modified manually. Resources such as gates, stands, and baggage
reclaims are unlikely to be allocated efficiently by time-consuming manual planning
processes. This results in delays for flights, delays for passengers, and increased
airport costs.
   On the effective date and time of the schedule, party role, transport, equipment,
and facility resources are allocated to the work. A system that automatically transfers
and integrates on-time data to automate the plan update is critical. Automating the
planning processes would reduce the time required to assess the impact of changes
from hours to minutes. The corresponding impact would be reduced delays, reduced
queues, and reduced costs.
208                                                                   K. Alexandridis et al.
Planning closely relates to the operational and functional cycles at an airport. Such
operational levels are tied to the airport’s physical and virtual or computational infras-
tructure functions. Figure 3 provides an overview of the proposed ACRIS Semantic
Model [44] integration with the coupled Digital Twin and agent-based modeling
technologies. The three basic groups of category package modules of the seman-
tic model (namely entities, moments, and motivations) are closely related to agent
classes (as entities), to agent actions (as moments), and agent belief-desire-intention
(BDI) goal architecture [45, 46] (as motivations). Therefore, according to Fig. 3, a
Digital Twin for a Smart Airport can be enriched by the standard semantic models
(ACRIS) coupled with agents’ behavior to realistically reflect the airport’s operation.
    More specifically, the knowledge organization of the ACRIS Semantic Model
integrates a number of operational modules, packages and data elements at the air-
port operational level [47] as can be seen in the following Fig. 4. Each of the three
fundamental objects (entities, moments, motivations), is mapped to each of the ele-
ment object library modules shown on the right side of Fig. 4. Each of these libraries
in turn has a number of nested attributes, operational layers and data associated
with them, and can be linked to the resources available to multi-agent BDI goals,
strategies, decisions, and actions.
    At the visual, 3D, and metaverse levels of reality capture, these operational ele-
ments of the DT-ABM-ACRIS integrative architecture can be visually, geospatially,
and locationally oriented in virtual metaverse space. The following images in Fig. 5
demonstrate an example of the reality capture process of the John Wayne Airport
(JWA) terminal in Orange County, California. The combination of stationary, mobile,
and aerial LiDaR scanning technologies allows us to fully capture and closely asso-
ciate centimeter-accuracy locational attributes of captured scenes, 3D objects and
point-cloud coordinates with the developmental aspects of the agent-based model-
ing and simulation process.
Distributed AI Modeling and Simulation for Smart Airport …                                 209
Fig. 3 Digital Twin and agent-based model integration with basic module package organization of
the ACRIS semantic model
2.4 Definitions
2.4.1 Allocation
Business rules are used to define the work to be done, the default work criteria, the
default resource types that are required to do the work, and the dependencies between
the work units. For example, the work unit “Land Aircraft” applies to all aircraft types
within the International Air Transport Association (IATA) aircraft type codes and will
require the default runway facility resource and also the participation of the air traffic
authority party role resource. The work unit “Taxi Aircraft to Stand” depends on the
210                                                                   K. Alexandridis et al.
Fig. 4 Package module and design organization of the modified ACRIS semantic model
“Land Aircraft” work unit having been successfully completed. This will apply to all
aircraft types and will require a taxiway and air traffic authority as default resources
(see example in Fig. 6).
Default Resource Criteria describe what a party wants to use resources for, such as
equipment or facility resources. Resource criteria define what a particular facility or
resource is required to be used for.
Resource constraints (usage constraints) specify the kind of work that a resource
cannot be used, for example, where an explosive detection system cannot be used
for Level 3 hold baggage screening. This is the opposite of resource usage criteria.
Distributed AI Modeling and Simulation for Smart Airport …                                    211
Fig. 5 Infrastructure operational visualizations of the John Wayne Airport (JWA) in Orange County,
California. Top Row: a, b LiDAR point-cloud and 360.◦ photosphere reality capture; c geodetic
monument examples at and around John Wayne Airport; d JWA monuments and high spatial
accuracy geodetic control network; e Orange County GPS real-time network ties
Fig. 6 Example inheritance and dependency structure of airport operation business rules
212                                                                 K. Alexandridis et al.
    Resource Usage Criteria, or just Usage Criteria, define what a particular facility
or resource can be used for. This is the opposite of resource (usage) constraints.
Resource usage criteria implicitly provide constraints on resources, for example
standard check-in desks can deal with standard bags, but not Out Of Gauge (OOG)
bags. An OOG desk can deal with standard bags as well as OOG bags, though at a
reduced throughput in terms of bags per minute.
    Usage criteria and capacities are provided to the planner by demand and capac-
ity management and they may differ for each environmental scenario category. An
example of a set of resource usage criteria and their characteristic properties is shown
in Table 3.
Work is a combination of work criteria and the resource that is to be applied to the
work. It typically comprises a hierarchy of units of work. For example the work to
land an aircraft includes taxiing the aircraft to the stand and parking it at the stand.
Work may have dependencies, e.g., aircraft separation must be maintained to 5 min
for code F, and may have a specified order in which it needs to be done, e.g., unload
baggage from aircraft and then move Unit Load Devices (ULDs) to arrivals break.
   Work Criteria is a standard for defining work to be done, including sufficient
information so that the work can be:
• Scheduled with competent resources.
• Readily determined to have been completed or not.
• Objectively measured.
Distributed AI Modeling and Simulation for Smart Airport …                          213
    Work criteria identify the type of work to be done, e.g., (1) unload baggage from a
British Airways 744 aircraft at 08:00 every Monday during the 2023 summer season,
or (2) sort and load bags for business class passengers traveling on a British Airways
flight to Boston. See Fig. 7 for further examples in the context of the required work.
    Work progress is the quantifiable completion of work. The actual doing of the work
is part of logistics and conformance. Monitoring the work progress by resources and
assets is part of performance management.
    In the plan model, a plan is a forward-looking work schedule. A work schedule
is an association of facility, equipment, and party role resources to units of work in
specified time periods. It takes business requirements into account, as provided in a
load-balanced capacity profile for a defined scenario. It also incorporates the usage
constraints of the resources that are to be included in the work schedule. Units of work
should have a global unique identifier to enable accurate allocation and deallocation/
reallocation of resources to the work.
Fig. 7 Arrival time probability distribution segmentation for incoming (left subgraph) and outgoing
(right subgraph) passengers of a single flight in an airport terminal
    Adopting such a distributed agent typology schema enables the introduction and
use of an agent orchestration framework within a broader BDI agent architecture
[45, 51]. Specifically,
Passenger-Agent: This archetypal agent typology captures and monitors basic
passenger-agent spatial movements across different airport and terminal sections.
It begins with the passenger transportation mode section by tracking the mode of
transportation (e.g., public transport or service transportation) along with parking
and parking spot allocation. Once passengers move into the main airport terminal
(e.g., airline check-in section and location), it tracks flight assignments, thus associ-
ating each passenger agent with a specific flight module (consequently, linking the
passenger-agent module with the flight-agent and terminal-agent modules). Further,
along the passenger journey through the airport, the passenger-agents relate their
location and assignment through the security check-in counters/locations, the choice
of vendors, and vendor location for airport concessions (thus associating passenger-
agent entities with vendor-agent entities). Finally, it tracks passenger movement
through their departure/arrival gates and embarkation/disembarkation queues.
Flight-Agent: This typology provides a distributed agent-based framework for all
flight-related components or modules related to smart airport operations. These
include the check-in management process (from the airline and provider operational
perspective), gate assignment for flights departing or arriving, time and process man-
agement for departures and arrivals, including delay and queue management, plane
movement within an airport (e.g., from/to gates, etc.), plane turnaround and airport
collaborative decision-making processes (A-CDM), as well a framework for modu-
larity interfacing agent roles with airline operational systems.
Terminal-Agent: The third typology relates to smart airport terminal-based oper-
ations. These agents track and distribute baggage movements, gate management
assignments (from an airport facility operational perspective), facility maintenance
and integrated management (in terms of both temporal and spatial perspectives), air-
216                                                                 K. Alexandridis et al.
port security (physical and network), and smart asset management. These terminal-
agent types interface directly with relevant system modules and provide distributed
intelligent communication interfaces with physical system modules.
Vendor-Agent: This agent typology relates to distributed communication and action
mechanisms related to vendor assignments, operations, product and service delivery
mechanisms, concession property and title management, and interface (externally)
with physical and virtual point-of-sale interfaces within an airport smart operations
system.
   Collectively, these examples of smart airport agent typologies are capable of pro-
viding the necessary role assignments and functions that enable a smart, robust, adap-
tive, and intelligent framework for agent actions, resource (management/allocation/
distribution), time and space interactions with physical movement, behavioral and
cognitive response mechanisms, and property and facility management functions.
While not necessarily exhaustive (in fact, alternative and augmented additional agent
types and typology groups can be designed and used), they form a cohesive and tightly
coupled set of agent roles that, through their exchanges and interactions, can give
rise to the emergence of complex and intelligent distributed system behaviors and
operational mechanisms.
This section focuses on the visual metaverse aspects of an intelligent Digital Twin
framework. It is mainly based on an award-winning TM Forum Moonshot Catalyst
project (URN M23.0.567) [13]. The project features innovative and active public-
private partnerships across a range of industry consortia, public organizations, and
airports. The remaining section content provides a short and concise outline of the
reference architecture and design as well as the structural and operational intelligence
modeling and simulation elements used in enhancing a Digital Twin coupled with
metaverse experience, supporting the implementation of the agent-based intelligent
ontology semantic planning framework described in the previous sections.
    Figure 9 provides an overview of the key elements of the interoperable airport
ecosystem for a metaverse approach. The elements range from those of high abstrac-
tion/fundamental composability (e.g., public transport, airports, airlines, passengers,
agencies), to more specific and metaverse domain-explicit concepts (retail, telecom,
AECO, traffic management, etc.). The intent-based service composition and orches-
tration, along with information exchange and cross-domain interoperability, are gov-
erned by a common and shared ontology framework, such as the ACRIS semantic
framework described in Sect. 3.
    The relevant reference implementation outline example is shown in Fig. 10. As the
back-end metaverse framework interacts with the entities, moments, and motivators
(e.g., person recognition and hazard recognition service layers), the mid-end layers
provide a design and reference orchestration for services like application communi-
cation, scenario modeling and simulation, data transformation and interoperability,
Distributed AI Modeling and Simulation for Smart Airport …                                217
Fig. 9 Elements of the interoperable airport ecosystem metaverse approach. The scalable archi-
tecture includes and encapsulates a common semantic ontology for cross-domain interoperability
Fig. 10 Real-time information exchange components and layers for metaverse reference imple-
mentation for the Digital Twin smart airport model
Fig. 11 Metaverse visualization and real-time Digital Twin dashboard operations: a a visual display
of navigational agent spatial movements; b visual check-point operations and service message
exchange real-time operations; c showcase of interactive airport parking allocation and navigation;
d passenger-based dashboard for real-time event-driven scenario simulations
services, such as travel management and product fulfillment-all the service orches-
tration and coordination/orchestration operations function within a near-real-time,
high-fidelity functional metaverse environment.
    In terms of the metaverse visualization experience (see Fig. 11), the nature and
characteristics of Digital Twin 3D immersion allow for both enhancing the visual
intelligence and locational/spatial navigational experience and the near real-time ser-
vice, data, and communication provision. While the former enhances and supports
visual intelligence operations, the latter promotes and encapsulates implementation
based on high accuracy, fidelity, and interoperability. Taken together, these two fun-
damental principles (visual intelligence and fidelity), when applied within a Digital
Twin metaverse model, work complementary to each other in establishing and bench-
marking a conceptual framework for intelligent Digital Twins.
    The overview of the overall smart airport and intelligent Digital Twin metaverse
implementation is provided in Fig. 12. The approach focuses on predictable, accurate,
and visually immersed outcomes stemming from automation, intelligent decision-
making, and inter-agent distributed intelligence. Both internally (within events, ser-
vice provisioning, resource allocation components, and orchestration mechanisms)
and externally (inputs and outputs such as data and services), it is essential to strive
to achieve a high degree of cross-standards interoperability and composability.
    Some of the benefits of integrative metaverse design, control, and operations
include the ability of airport management and operational planning capable to sim-
ulate and experimenting with alternative scenarios before implementation, reducing
Distributed AI Modeling and Simulation for Smart Airport …                           219
cost (design and implementation) while managing risk and risk expectations (includ-
ing risk aversion and risk propensity attitudinal dispositions). As a result, an intelli-
gent metaverse Digital Twin framework streamlines time-to-value operations while,
in the near- and long-term, promoting higher return-on-investment ratios. At the
same time, the Digital Twin framework enables the development and establishment
of broader sustainability and robustness characteristics at the airport level. These
might include emergency and incident response or reporting operations, addressing
sustainability goals and criteria, and forming a more innovation-minded workforce.
In this chapter, we proposed a new generation of ABM that leverages the metaverse
and Digital Twin capabilities. Also, we highlighted the significant role of real-time
data for a more realistic representation of Airport operational activities and planning
processes. The aim is to optimize and enable airport planning with predictive analysis
to understand and adapt to unpredictable issues. We argued that the complexity of
airport systems cannot be addressed by linear and ad hoc modeling and simulation
methods. To address this challenge, we described two methods of categorizing ABMs
to set up both the use of phenomena and exploratory-based models in the airport
planning domain and introduce agent and communication standards. It was shown
220                                                                          K. Alexandridis et al.
that many agent systems use the standard Foundation for Intelligent Physical Agents
and Agent Communication Language (FIPA-ACL) to support the sharing of message
content with common semantics that includes communication context. This context
explicitly defined relationships between a particular message in relation to a context
such as a sender agent’s current workflow, an agreed interaction protocol, goals, and
plans, or with respect to a receiving agent’s beliefs, desires, and intentions [53]. The
OMG Agent and Event Metamodel [31] standard will supersede FIPA-ACL but is
planning to include ACL compliance. Future work will trade-off ACL working with
the message exchange formats and transport/application protocols.
   We provided a key perspective to airport planning model development, including
an ecosystem for context planning, a planning cycle with a workflow standard model,
integration of the Digital Twin, metaverse, and ABM model, and key definitions.
   The provided conceptual model has practical and theoretical implications. From a
practical perspective, different operational activities can benefit the role of Metaverse
and Digital Twins to test the human-environment interactions in various scenarios.
This capability can be used for the interaction of different stakeholders, such as air-
lines, third-party services, and safety and security teams. The agent-based modeling
and simulation, calibrated with real-time data, is a new line of research that needs
further exploration to offer a reliable and trustworthy output for decision-makers
in airport operations. In addition, an ontological framework is critical to address
the interoperability and harmonization of complex data and analytics in an airport
ecosystem.
   Finally, this document provided an agenda for developing a proof of concept
and pilot study to be implemented in an actual case and an examination of the
idea of a new generation of ABM integrated into the Digital Twin architecture and
detailed information and guidelines for Airport Planning, which can be used for
ABM development.
Acknowledgements The work presented in this chapter draws from a fertile collaborative environ-
ment for discussions, discourse, and collaboration of a number of collective organizations. Digital
Twin Consortium (DTC) working groups are one of these organizations. S. Sabri and K. Alexan-
dridis serve as co-chairs of the DTC Academia and Research Working Group, while D. Migliori
serves as one of the co-chairs of the DTC Mobility and Transportation Working Group. Many of
these ideas and concepts emerged through its collaborative environment and the author’s interac-
tions within it. The parent industry standards development organization Object Management Group
(OMG) is another key organization we wish to acknowledge. Jeff Smith serves as co-chair of the
OMG Analysis and Design Task Force and co-submitter of the AgEnt specification. TM Forum and
the Airports Council International (ACI) provided the necessary background for the development
of ideas, projects, and resources. Specifically, the metaverse implementation related to passenger
experiences, provided in Sect. 4, is an integral part of an award-winning Moonshot Catalyst program
of TM Forum and represents an active collaboration of many industry and public organizations.
D. Migliori, B. Logan, and K. Alexandridis serve as key organizational and industry champion
members of the project. Furthermore, the integration of the ACRIS Semantic Model described in
Sect. 3 owes its roots and fundamental concepts to the ACI’s Aviation Community Recommended
Information Services (ACRIS) model approach. B. Logan is one of the key architects and custo-
dians of the ACRIS semantic model organization and its resources. Finally, the authors wish to
acknowledge a number of individuals and organizational representatives for the exchange of ideas
and organizational support: Dan Isaacs, general manager of OMG and CTO of DTC for his sup-
Distributed AI Modeling and Simulation for Smart Airport …                                      221
port of DTC working groups and organizational facilitation; Members of various DTC working
groups for feedback and review of this work and ideas over various stages; Member-organizations
of the TM Forum Moonshot Catalyst Project No. M23.0.567 for supporting and recognizing the
transformative concepts.
References
13. TM Forum (2023) Transforming passenger experiences with continuous decision intelligence
    (Moonshot Catalyst URN M23.0.567)
14. Zhang H, He B, Lu G, Zhu Y (2022) A simulation and machine learning based optimization
    method for integrated pedestrian facilities planning and staff assignment problem in the multi-
    mode rail transit transfer station. In: Simulation modelling practice and theory, vol 115. 9
    citations (Crossref) [2023-09-25] Elsevier B.V
15. Peng J, Wei Z, Wang S, Qiu S (2023) Toward dynamic regulation of bidirectional automatic fare
    gates: a macro-to-micro simulation approach. In: Simulation modelling practice and theory,
    vol 124. 1 citations (Crossref) [2023-09-25] Elsevier B.V
16. Delcea C, Cotfas LA, Milne R, Xie N, Mierzwiak R (2022) Grey clustering of the variations in
    the back-to-front airplane boarding method considering COVID-19 flying restrictions. In: Grey
    systems, vol 12, no 1, pp 25–59. 8 citations (Crossref) [2023-09-25] Emerald Group Holdings
    Ltd
17. Zellers SJH, Bradford WC, Snowden CH, Knoedler RJ, Alexander JE, Coalson MB, Sonkin
    J, Foley C, Stinson W, Tansey P, Smith JF, Quinn J, Williams K, Whitfield-Smith L (2024)
    Airport energy resiliency roadmap. Transportation Research Board, Washington, DC
18. Xianliang G, Jingchao X, Zhiwen L, Jiaping L (2021) Analysis to energy consumption char-
    acteristics and influencing factors of terminal building based on airport operating data. In:
    Sustainable energy technologies and assessments, vol 44. 4 citations (Crossref) [2023-09-25]
    Elsevier Ltd
19. Gu X, Xie J, Huang C, Ma K, Liu J (2022) Prediction of the spatiotemporal passenger distribu-
    tion of a large airport terminal and its impact on energy simulation. In: Sustainable cities and
    society, vol 78, p 103619. 5 citations (Crossref) [2023-09-25]
20. Gu X, Xie J, Huang C, Liu J (2022) A spatiotemporal passenger distribution model for airport
    terminal energy simulation. In: Indoor and built environment, vol 31, pp 1834–1857. 1 citations
    (Crossref) [2023-09-25]
21. Li Z, Zhang J, Mu S (2023) Passenger spatiotemporal distribution prediction in airport terminals
    based on insect intelligent building architecture and its contribution to fresh air energy saving.
    In: Sustainable cities and society, vol 97, p 104772. 0 citations (Crossref) [2023-09-25]
22. Lin L, Liu X, Liu X, Zhang T, Cao Y (2023) A prediction model to forecast passenger flow
    based on flight arrangement in airport terminals. In: Energy and built environment, vol 4, no
    6, pp 680–688. 1 citations (Crossref) [2023-09-25] KeAi Communications Co
23. Janssen S, Sharpanskykh A, Curran R (2019) Agent-based modelling and analysis of security
    and efficiency in airport terminals. In: Transportation research part C: emerging technologies,
    vol 100, pp 142–160. 18 citations (Crossref) [2023-12-04]
24. Zhang X, Chen G, Shi J (203) Simulation of guided crowd evacuation scheme of high-speed
    train carriage. Int J Simul Modell 22(1):110–120. 0 citations (Crossref) [2023-09-25] DAAAM
    International Vienna
25. Lahoti N (2022) How the metaverse will change transportation and future mobility
26. Darvishi M, Ahmadi G (2014) Validation techniques of agent based modelling for geospatial
    simulations. Int Arch Photogrammetry Remote Sens Spatial Inf Sci XL-2-W3:91–95. 8 citations
    (Crossref) [2023-12-04] Conference Name: WG II/1, WG II/4, ICWG II/IV, WG IV/7, The
    1st ISPRS International Conference on Geospatial Information Research (Volume XL-2/W3)
    - 15–17 November 2014, Tehran, Iran Publisher: Copernicus GmbH
27. Smaldino P (2023) Modeling social behavior: mathematical and agent-based models of
    social dynamics and cultural evolution. Princeton University Press. Google-Books-ID:
    aOS3EAAAQBAJ
28. Scala P, Mota M, Blasco-Puyuelo J, Garcia-Cantu O, Blasco C (2022) A novel validation
    approach for validating the simulation model of a passengers’ airport terminal: case study Palma
    de Mallorca airport. In: European modeling and simulation symposium, EMSS. 0 citations
    (Crossref) [2023-09-25]
29. Wilensky U, Rand W (2015) An introduction to agent-based modeling: modeling natural,
    social, and engineered complex systems with NetLogo. MIT Press, Google-Books-ID, LQrhB-
    wAAQBAJ
Distributed AI Modeling and Simulation for Smart Airport …                                        223
30. Carbo J, Sanchez-Pi N, Molina JM (2018) Agent-based simulation with NetLogo to evaluate
    ambient intelligence scenarios. J Simul 12:42–52. 9 citations (Crossref) [2023-09-25] Pub-
    lisher: Taylor & Francis _eprint: https://doi.org/10.1057/jos.2016.10 Northwestern: https://
    ccl.northwestern.edu/2016/carbo.pdf
31. OMG (2022) Agent and Event Metamodel (AgEnt) joint revised submission. "AgEnt" Team
    0.4-01, Object Management Group
32. Vlăduţescu, Stănescu G (2023) Environmental sustainability of metaverse: perspectives from
    Romanian developers. Sustainability (Switzerland) 15(15)
33. Khalaj O, Jamshidi M, Hassas P, Mašek B, Stadler C, Svoboda J (2023) Digital twinning of
    a magnetic forging holder to enhance productivity for industry 4.0 and metaverse. Processes
    11(6)
34. Ma Z (2023) Energy metaverse: the conceptual framework with a review of the state-of-the-art
    methods and technologies. Energy Inf 6(1)
35. Negi P, Singh R, Gehlot A, Kathuria S, Thakur A, Gupta L, Abbas M (2023) Specific soft com-
    puting strategies for the digitalization of infrastructure and its sustainability: a comprehensive
    analysis. Arch Comput Methods Eng
36. Naderi H, Shojaei A (2023) Digital twinning of civil infrastructures: current state of model
    architectures, interoperability solutions, and future prospects. Autom Constr 149
37. Pamucar D, Deveci M, Gokasar I, Tavana M, Köppen M (2022) A metaverse assessment
    model for sustainable transportation using ordinal priority approach and Aczel-Alsina norms.
    In: Technological forecasting and social change, vol 182
38. Deveci M, Pamucar D,Gokasar I, Martinez L, Köppen M, Pedrycz W (2024) Accelerating the
    integration of the metaverse into urban transportation using fuzzy trigonometric based decision
    making. In: Engineering applications of artificial intelligence, vol 127
39. Grupac M, Negoianu A-E (2023) Immersive extended reality and sensor-based object recogni-
    tion technologies, socially-oriented location tracking and simulation modeling tools, and arti-
    ficial vision and haptic augmented reality systems in the metaverse interactive environment.
    Rev Contemporary Philos 22:226–243
40. Grieves M, Hua E (2024) Defining, exploring, and simulating the Digital Twin metaverses. In:
    Digital Twins, simulation, and metaverse: driving efficiency and effectiveness in the physical
    world through simulation in the virtual worlds, Forthcoming. Springer, Heidelberg
41. Xu Z, Bai Q, Shao Y, Hu A, Dong Z (2022) A review on passenger emergency evacuation from
    multimodal transportation hubs. J Traffic Transp Eng (English Edition) 9:591–607. 2 citations
    (Crossref) [2023-09-25]
42. Wooldridge M, Jennings NR (1995) Intelligent agents: theory and practice. In: The knowledge
    engineering review, vol 10, pp 115–152. 3452 citations (Crossref) [2023-09-25] Cambridge
    University Press
43. Benfield SS, Hendrickson J, Galanti D (2006) Making a strong business case for multiagent
    technology. In: Proceedings of the fifth international joint conference on autonomous agents and
    multiagent systems, (Hakodate Japan). ACM, pp 10–15. 21 citations (Crossref) [2023-09-25]
44. Airports Council International (2019) ACI-ACRIS semantic model of SOA, tech rep, Airports
    Council International (ACI)
45. Grimm V, Berger U, Bastiansen F, Eliassen S, Ginot V, Giske J, Goss-Custard J, Grand T,
    Heinz SK, Huse G, Huth A, Jepsen JU, Jorgensen C, Mooij WM, Muller B, Pe’er G, Piou
    C, Railsback SF, Robbins AM, Robbins MM, Rossmanith E, Ruger N, Strand E, Souissi S,
    Stillman RA, Vabo R, Visser U, DeAngelis DL (2006) A standard protocol for describing
    individual-based and agent-based models. Ecol Modelling 198(1–2):115–126. 1896 citations
    (Crossref) [2023-09-25]
46. Alexandridis KT (2006) Exploring complex dynamics in multi agent-based intelligent systems:
    theoretical and experimental approaches using the Multi Agent-based Behavioral Economic
    Landscape (MABEL) model. Ph.D. Dissertation, Purdue University, West Lafayette, Indiana,
    May 2006. ISBN: 9780542864230 Publication Title: Forestry & Natural Resources Volume:
    Doctor of Philosophy 3232142
224                                                                           K. Alexandridis et al.
47. Alayande S, Airports Council International (2017) ACRIS semantic model guide to the knowl-
    edge layer. Technical Report 1.1, Airports Council International (ACI)
48. Tenkanen H, Toivonen T (2020) Longitudinal spatial dataset on travel times and distances by
    different travel modes in Helsinki Region. Sci Data 7:77. 16 citations (Crossref) [2023-09-25]
    Number: 1. Nature Publishing Group
49. Agresti A (2003) Introduction: distributions and inference for categorical data. In: Categorical
    data analysis, Wiley Series in Probability and Statistics. John Wiley & Sons - Wiley Inter-
    Science, 2nd edn, pp 1–35, Section: 1
50. Crane R, Sornette D (2008) Robust dynamic classes revealed by measuring the response func-
    tion of a social system. In: Proceedings of the national academy of sciences, vol 105, pp
    15649–15653. 535 citations (Crossref) [2023-09-25]
51. Wai Sk, WaiShiang C, Khairuddin MAB, Bujang YRB, Hidayat R, Paschal CH (2021)
    Autonomous agents in 3D crowd simulation through BDI architecture. JOIV : Int J Inf Visu-
    alization 5:1–7. 2 citations (Crossref) [2023-12-04] Number: 1
52. Palanca J, Rincon JA, Carrascosa C, Julian VJ, Terrasa A (2023) Flexible agent architecture:
    mixing reactive and deliberative behaviors in SPADE. Electronics 12:659. 1 citations (Crossref)
    [2023-12-04] Number: 3. Multidisciplinary Digital Publishing Institute
53. Poslad S (2007) Specifying protocols for multi-agent systems interaction. ACM Trans
    Autonomous Adaptive Syst 2:15–es. 94 citations (Crossref) [2023-12-04]
The First Real-Time Digital Twin
of a Nuclear Reactor
Abstract In this chapter we examine the first fully operation digital twin of a fissile
nuclear reactor. In 2023, Idaho National Laboratory (INL) partnered with Idaho State
University (ISU) to create and run a digital twin of their AGN-201 fissile nuclear
reactor. This reactor can produce up to 5 watts and has proven to be the perfect
testbed for discovering how best to implement a digital twin of a larger reactor
system. INL created an integrated, cloud-based digital twin capable of measuring
data from the reactor in near real-time and built the foundation for running operations
such as machine learning and analytics in near real-time in an offsite location. This
experiment is a key step in building a digital twin of a larger reactor system and has
helped highlight many potential pitfalls and problems that such an endeavor might
face. This experiment has also shown the great promise that a cloud-first approach
has when creating digital twins.
The Idaho National Laboratory (INL), established post-World War II, has played a
significant role in the development of nuclear technology, and now Digital Twins.
Initially named the National Reactor Testing Station, it was the site where, in 1951,
the Experimental Breeder Reactor I (EBR-I) first generated electricity from nuclear
power.
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2024           225
M. Grieves and E. Hua (eds.), Digital Twins, Simulation, and the Metaverse, Simulation
Foundations, Methods and Applications, https://doi.org/10.1007/978-3-031-69107-2_10
226                                                                     J. Darrington et al.
    During the 1960s and 1970s, INL focused on researching advanced reactor designs
and improving nuclear safety. By the 1980s, the laboratory shifted its attention to
environmental management, particularly the remediation of nuclear waste.
    In recent decades, INL has continued to work on nuclear reactor technology,
emphasizing extending reactor lifespans, advancing next-generation reactors, and
contributing to small modular reactor (SMR) research. It also engages in cyber-
security for nuclear facilities and supports nuclear nonproliferation and security
initiatives.
    Today, INL is involved in developing nuclear energy technologies with an
increased emphasis on addressing climate change. Throughout its history, INL has
had a substantial influence on the evolution of nuclear energy policy and infrastructure
in the United States.
    For the past 3–5 years, the Idaho National Laboratory (INL) has been championing
the use and creation of DTs in the nuclear industry. This consists of developing virtual
DTs to evaluate potential reactor misuse, and for diversion scenarios for sodium fast
reactors and high-temperature Pebble Bed Reactors [1].
    According to the United States Department of Defense, Digital Engineering is
a comprehensive approach that relies on authoritative sources of system data and
models, seamlessly spanning multiple disciplines to support activities throughout
a product’s entire lifecycle, from conceptualization to disposal [2]. Unlike tradi-
tional static documents, digital engineering emphasizes the use of interconnected and
dynamic models to describe a product. These models are integrated across diverse
platforms to facilitate the development of a design product that can be effectively
supported throughout its operational life [3, 4].
    Digital Twins leverage the principles of digital engineering to create a model
that represents either an existing physical product or the design of a future physical
product [4]. This concept allows for the creation of two distinct types of Digital Twins:
a virtual Digital Twin, which simulates the behavior of a physical product through
modeling and simulation, and a Digital Twin of a physical system, which combines
a physical asset with computational models to provide a virtual representation of the
asset.
The adoption of Digital Twins has begun to revolutionize various aspects of nuclear
plant operations, maintenance, and training [1, 5, 6]. Plant reference simulators serve
as a rudimentary form of Digital Twins, providing operators with a training tool.
However, these simulators are not directly connected to the actual power plant.
Furthermore, Digital Twins are increasingly being applied in the realm of condi-
tion monitoring. Online performance and condition data can be collected through
sensors, which are then processed by various artificial intelligence and machine
learning algorithms to provide valuable insights.
The First Real-Time Digital Twin of a Nuclear Reactor                               227
establish a template for data-driven evaluations that are instrumental in the creation
and oversight of emerging aqueous metal separations technologies.
2.2 Design
DTs offer transformative potential for nuclear reactor design, particularly during the
basic and final stages. These sophisticated virtual models can be instrumental in eval-
uating both the reactor’s nuclear components for safeguards and the Nuclear Power
Plant’s (NPP) overall security. A pivotal role of a DT at this stage is to ensure the
design incorporates the necessary infrastructure for security and safeguards moni-
toring, including strategic placement of sensors within the reactor core and facility
to maintain operational knowledge.
2.3 Security
The AGN-201 is a compact nuclear reactor with a power output of 5.0 watts, situated
at Idaho State University. This reactor has been designed for both research purposes
and practical training [8]. Its structure comprises two primary components: the core
region and the ex-core elements.
    The core region encompasses nine fuel disks that consist of a homogeneous
mixture of polyethylene and uranium dioxide fuel, enriched uranium. The core’s
approximate dimensions are 24 cm in height and 25.6 cm in diameter [68]. The core
is divided into two sections connected by a thermal fuse made of polystyrene, which
contains twice the fuel loading of the fuel disks. The thermal fuse serves as a safety
feature, it will melt and separate the core into two pieces, thereby terminating the
chain reaction if the temperature rises above a specified threshold.
    To sustain criticality, four fuel control rods are inserted into the core. Two safety
rods remain fully always inserted, while a coarse control rod (CCR) and a fine control
rod (FCR) are employed to adjust, reduce, or maintain the reactor’s power output.
Additionally, there is a central irradiation facility passing through the core’s center,
allowing for experiments to be directly inserted into the core.
    The core is enclosed by a graphite reflector contained within an aluminum core
tank. Surrounding this tank is an additional layer of graphite reflector, followed by a
lead shield, and finally, a water shield. Above the core, there is a removable graphite
thermal column that can be taken out for experimental purposes, and below the core
are the control rod drive mechanisms. The complete configuration of the AGN-201
reactor is illustrated in Fig. 1.
There are numerous advantages of using DT technologies. They contain the ability to
encompass many of the necessary international safeguards and security of reactors,
but there are also challenges that arise when implementing a DT. Using any digital
product such as a Digital Twin, artificial intelligence, and machine learning, risk and
uncertainty are introduced into the system, and can possibly become a high risk.
Some of the challenges of using these products are:
1. Absence of regulatory guidance and requirements for compliance and acceptance
   of use.
230                                                                  J. Darrington et al.
5 Technical Implementation
Digital Engineering and the development of digital twins necessitate the use of highly
sophisticated tools and software [26]. DTS allow the early detection, deterrence, and
response to potential threats by leveraging AI/ML algorithms for real-time moni-
toring of reactor data streams. Idaho State University and the Idaho National Lab
partnered up to create a digital twin centered around the AGN-201 reactor.
The First Real-Time Digital Twin of a Nuclear Reactor                              231
5.1 Jester
1   https://github.com/idaholab/Deep-Lynx.
232                                                                              J. Darrington et al.
5.2 DeepLynx
Fig. 2 Data flows into the data acquisition system from the reactor. Jester is used to push data from
the DAS into a DeepLynx cloud instance. The data is analyzed using machine learning algorithms
implemented through Juypter notebooks managed through Papermill, and pushes those results back
into DeepLynx, which then can be viewed by users through DeepLynx’s UI or the operator Window’s
program
The First Real-Time Digital Twin of a Nuclear Reactor                                     233
Fig. 3 DeepLynx representation of the AGN-201 in graph format. This representation of the data
allows a human readable view of the time-series data collected from the AGN-201 reactor
data mining to obtain patterns or models from the gathered data from all kinds of envi-
ronments [25]. It has the power that relational database management systems have
with the bonus of object-oriented features such as defining complex data types, over-
loading functions to work with different argument data types, and defining inheritance
relationships between tables.
Timeseries Data
DeepLynx is designed to not only store a graph representation of the AGN-201
reactor, but also serves as the storage for the data read from the sensors themselves,
as previously stated. Currently, DeepLynx leverages TimescaleDB for its storage
and querying of time-series or tabular data. This is another open-source technology
which was adopted by the DeepLynx team.
   Data is processed only in near real-time due to network connectivity via HTTPS.
The different components of Fig. 1 are connected via TCP sockets using the HTTPS
format which is not necessarily the best for high throughput data. This stops the
system from producing real-time results. Currently the system runs in about thirty
second loops, including the machine learning and AI portions of the flow (Figs. 4
and 5).
234                                                                         J. Darrington et al.
We have made significant progress in realm of digital twin visualization for the AGN-
201 project. We have developed programs using various tools to enable operators
and monitors to easily identify anomalies and issues in reactor operation.
The First Real-Time Digital Twin of a Nuclear Reactor                                         235
Fig. 6 The operator UI allows operators to view trends and monitor the reactor in a human readable
format. This is a screenshot of the Windows desktop application
Operator UI
The operator UI is a beta desktop application designed to give insights into reactor
function to a reactor operator and monitor. This system incorporates the machine
learning data ingested by DeepLynx from the program listed previously and displays
it to the end user in an easy to digest format. This desktop program will allow
operators to digest the data from the prediction models and use that information to
understand risks and problems in the reactor proactively. The operator UI was created
by using a combination of Tauri,2 an open-source tool that builds optimized, secure,
and frontend-independent application for multi-platform deployment, and Svelte,3
an open-source front end component framework/language (Figs. 2 and 6).
2   https://tauri.app/.
3   https://svelte.dev/.
236                                                                     J. Darrington et al.
A key aspect of the AGN-201 digital twin is the machine learning and artificial
intelligence detection programs. These programs take the live data from the reactor,
pulling it directly from DeepLynx, and performs various tasks to determine different
things. We will discuss the actual models and mathematics later on this chapter, for
now we will focus only on the technology involved in the operations.
Papermill
Papermill4 is an open-source tool that allows you to execute Jupyter Notebooks an
parameterize them. This allows the execution of workflows without manually having
to combine different notebooks. INL-developed a tool in Rust to manage Papermill’s
execution of various Jupyter notebooks (Fig. 1, Sect. 2.1.1). This allows the system
to use the memory efficiency/security of Rust, while leveraging the power/flexibility
of Python for machine learning. This allows us to maintain a highly available, highly
resilient system while at the same time letting data scientists work in a familiar system
and paradigm. This system is contained in a soon to be open-sourced Docker5 image.
DuckDB
DuckDB6 is an in-memory process/in process SQL OLAP database management
system. It leverages columnar storage which makes it ideal for analytics and allows a
simpler data format that data scientists can read. We utilize DuckDB as a temporary
store for the data from DeepLynx, putting it into an easily queried and read paradigm
that allows data scientists to focus on their algorithms instead of the data fetching
and network concerns of communicating with DeepLynx.
4 https://papermill.readthedocs.io/.
5 https://www.docker.com/.
6 https://duckdb.org/
The First Real-Time Digital Twin of a Nuclear Reactor                                  237
5.5 Conclusion
INL and ISU was able to build a DT with near real-time system for anomaly predic-
tion and general analysis of the AGN-201 reactor. There were several lessons learned
from this process that will continue to advance the understanding of how DTs are
implemented. First, the DT is near real-time due to the limitations of data trans-
mission through https and security firewalls. This constrained the DT to leverage
batch processing instead of streaming the data from the AGN-201 reactor. Second,
rust is a newer language and needed a lot of debugging to establish a dependable
interop between rust and python to effectively manage and analyze reactor data being
processed by Deep Lynx.
The model framework consists of two primary components: (1) Model Development
and (2) Model Implementation and the models attempt to predict the following: Ch2
(Watts), Ch3 (Watts), FCR (cm), and CCR (cm).
   Our methodology in machine learning (ML) serves as an automated system for
quality assurance and control (QAQC), as well as a framework for probing different
methods. The core of our strategy is the formulation of models that predict outcomes
for critical components such as Coarse Control Rods (CCR), Fine Control Rod
Heights (FCR), and the power outputs of Channel 2 and Channel 3 (measured in
238                                                                     J. Darrington et al.
Fig. 7 Flow of data from the AGN-201 through the mathematical SM and PKE-SM
Watts). We achieve this by employing all available variables, excluding the target
variable, to serve as predictors within our models. The predictor variables, detailed
in Table 1, include Channel 1 count rate (Ch1, CPS), power readings from Channel
2 (Ch2, W) and Channel 3 (Ch3, W), positional measurements of Fine Control
Rods (FCR, cm) and Coarse Control Rods (CCR, cm), Temperature (Temp, °C), and
Inverse Period (Inv-P, sec).
    In this research, we have adopted three distinct model architectures, namely Multi-
layer Perceptron (MLP) Regression, GrowNet, and TabNet, each bringing a unique
perspective to the analysis of tabular data.
    Multilayer Perceptron Regression, drawing upon classical neural network
paradigms, serves as a foundational model to set a benchmark for performance
comparison [9]. On the other hand, GrowNet is a novel gradient-boosted approach
that employs a composite of simpler neural networks, which are iteratively combined
to form an intricate and robust model adept at optimizing a variety of loss functions
[10]. A notable feature of GrowNet is its comprehensive corrective phase, which
fine-tunes the parameters across all contributing networks to enhance the model’s
predictive capacity.
    TabNet, employing a sequential attention mechanism within a deep learning
context, stands out by sequentially parsing through variables to identify and prioritize
the most significant ones in relation to the overall dataset features [11]. This focus
on salient variables allows TabNet to make more informed predictions by effectively
allocating computational attention where it’s most impactful.
    Together, these three methodologies represent the spectrum from tried-and-true
to state-of-the-art, encapsulating a broad range of capabilities in tabular data analysis
and providing a robust suite of tools for developing alongside digital twins.
The First Real-Time Digital Twin of a Nuclear Reactor                                 239
The model development portion of the framework consists of the following steps:
1. Training Data Preprocessing: The training data set is cleaned using a rule-based
   quality assurance/quality control (QAQC) along with a standard normalization.
   First, the data is scrubbed by reviewing the data for errors that violate the reactor’s
   known data thresholds, specifically identifying and fixing values that exceed
   power maximum, and fall underneath reactor power minimums. This also consists
   of populating any missing and incorrectly formatted missing data (NaN values).
   Each variable is then adjusted using a standard normalization function. Finally,
   the mean values and standard deviation values are stored for subsequent access.
2. Model Training: The training data is split into training, testing, and validation
   datasets. This function is serialized so that a unique model is developed for each
   variable.
3. Meta Data Storage: This step allows the model implementation step to access
   these values for consistent data normalization.
4. Trained Model Export: Each trained model is exported. The twin can accom-
   modate any file format that is appropriate for a specific model type.
The near real-time model implementation portion of the framework consists of the
following steps:
1. Data Ingestion: The framework incrementally ingests data via SQL queries,
   formats it to match the original training data, and stores the information in
   memory.
2. Data Preprocessing: The framework isolates each data time step individually
   and performs an automated QAQC and standard normalization using the mean
   and standard deviation values it retrieves from the meta data produced during
   model development.
3. Prediction: The trained model files are accessed and used to predict the values
   of each variable (FCR, CCR, water temperature, Ch1 power, Ch2 power, Ch3
   power) for the isolated and preprocessed data set.
4. Denormalization: The model predictions are returned to their original units by
   undoing the normalization function.
5. Result Storage: The model results include the reported and predicted values
   for each variable along with how those values differ. Under normal conditions
   while the reactor is critical, the reported and predicted values should be nearly
   identical. Deviation between those values is indicative that either (1) additional
   model training is required, (2) sensor failure, or (3) the reactor is not operating
240                                                                       J. Darrington et al.
      under expected conditions. The model does not attempt to identify which of those
      three scenarios is taking place, but the discrepancy is stored in the results so that
      it can be investigated further.
Creating a nuclear safeguards digital twin leverages the combined power of cloud
computing and high-performance computing (HPC). While the bulk of the digital
twin safeguards reside on the cloud, they collaborate with the INL HPC for opti-
mization algorithms and physics computations. This setup endows the digital twin
with the robust computational strength of HPCs and the flexibility and adaptability
of cloud storage.
   Simultaneously, a groundbreaking digital twin has been developed, tested, and
validated for the autonomous control of live assets within the Deep Lynx ecosystem.
This advancement, funded through the INL Laboratory Directed Research and Devel-
opment (LDRD) initiative, incorporated tools like Deep Lynx, the ML adapter,
and physics simulations grounded in the Multiphysics Object-Oriented Simulation
Environment (MOOSE). The INL’s Microreactor Agile Non-Nuclear Experimental
Testbed (MAGNET) serves as a testing platform for microreactor technologies.
Notably, on March 30, 2022, the LDRD digital twin was employed in a MAGNET
test involving a single heat pipe. The test showcased the digital twin’s prowess in
forecasting temperature benchmarks and adjusting MAGNET’s temperature set point
autonomously, enabled by the seamless two-way communication between Deep Lynx
and the MAGNET data acquisition system.
   The digital twin is slated for hosting on a blend of the Microsoft Azure cloud and
HPC infrastructure. Given the rigorous computational requirements of the Serpent
simulations, an HPC environment is best suited for rapid simulation processing.
While the Serpent simulation is contained within the INL HPC cluster, other aspects
of the digital twin will be housed in the Azure cloud. This arrangement offers
several advantages, such as scalability, cost-efficiency, enhanced security, and deeper
insights into hardware and network dynamics. For seamless data exchange between
the cloud and HPC, specific components and a data-flow framework are imperative.
For instance, Deep Lynx in the cloud can initiate an HPC processing request. An
HPC adapter fetches required files from Deep Lynx and secures them in the Azure
storage. Within the HPC, a service node constantly scans the Azure storage for fresh
requests and related files. Post-processing, results get transferred back to the cloud
storage. This HPC adapter consistently monitors the storage, relaying results back
to Deep Lynx once detected.
The First Real-Time Digital Twin of a Nuclear Reactor                               241
8 Visualization
QR code scanning plays a significant role in our workflow by enabling precise posi-
tioning of virtual assets relative to the physical world. This process allows engineers
to conduct design reviews, test ergonomics, fitment, escape routes, and address prac-
tical challenges, ultimately enhancing the design and testing phases (Idaho National
Laboratory, 2023).
242                                                                     J. Darrington et al.
Utilizing the game creation engine Unity (Unity Technologies n.d.), INL has devel-
oped a robust environment for visualizing digital twins. This platform was chosen
for its versatility in building projects across various platforms and its supportive
community.
   INL begins by converting Computer-Aided Design (CAD) models into a mesh
that is easier to render using PIXYZ, a tool designed to optimize models by simpli-
fying geometry without compromising on detail (Unity Technologies n.d.-a). The
primary advantage of integrating XR with data analysis lies in its capacity to facilitate
the interpretation and representation of intricate datasets. Conventional approaches
often rely on subject matter experts to manually collect, examine, and report anoma-
lous data. Our approach, however, offers spatial visualization capabilities, simpli-
fying the process of data identification and plotting. This real-time or near real-time
engagement enables operators to swiftly detect and understand unknown signals and
anomalies, elevating our decision-making capabilities.
   While HoloLens 2 remains our primary choice, the industry’s direction empha-
sizes cross-platform compatibility. OpenXR platform enables headset makers to
define hardware specifications while utilizing standardized interaction methods,
ensuring that applications can be developed for multiple headsets with ease (Idaho
National Laboratory 2023). Following the conversion process, these assets are
imported into Unity, where INL engineers craft interactive experiences tailored for
XR headsets like the HoloLens 2 (Microsoft n.d.), known for its Mixed Reality Toolkit
(MRTK) that facilitates the development of immersive applications with features
such as hand tracking, hand poses, speech recognition, eye tracking, gaze tracking,
speech recognition, and other interaction methods (Microsoft n.d.-a). Alternatively,
the industry is moving toward the OpenXR standard, which promises cross-platform
compatibility for XR applications (The Khronos Group n.d.).
The First Real-Time Digital Twin of a Nuclear Reactor                             243
An innovative feature utilized by INL is the Serval Multiplayer server, which enables
collaborative sessions by synchronizing state data among multiple users, even though
they operate in individual sessions. By allowing users to scan the same QR code, they
can view models in synchronized positions while maintaining individual sessions,
ensuring privacy. This enhances the collaborative process, allowing simultaneous
interaction with and visualization of the digital twin (Serval Project n.d.).
   In conclusion, INL’s integration of 3D visualization and XR technologies into
digital twins represents a significant advancement in the field, offering enhanced
data interpretation, real-time operational insights, and improved collaborative
capabilities.
Necessary Terminology
CCR         Coarse control rod.
DT          Digital twin.
FCR         Fine control rod.
IFML        Isolation forest machine learning.
INL         Idaho National Laboratory.
244                                                                     J. Darrington et al.
9.1 Overview
In the past year, the reactor operators at Idaho State University (ISU) operated the
AGN-201 reactor under various conditions without informing the digital twin (DT)
team at Idaho National Laboratory (INL) to perform both offline and online detec-
tion of hypothetical proliferation attempts. These data were transmitted to DeepLynx
during the operation and stored for analysis by the DT team. During this time, knowl-
edge of the actual operations was kept from the DT team to allow for an offline anal-
ysis and estimate of events. Three separate analyses (reactor physics assessment,
machine learning predictions, and anomaly detection) were performed by the DT
team to examine the DT’s ability to detect and quantify off-normal operations. This
provided first-of-a-kind online and offline monitoring of a nuclear reactor using a
DT, proves the impact, and necessity of the utilization of digital twin’s in the realm
of security, and validation of critical systems.
The red/blue team test involved the ISU reactor supervisor (red team) modifying
the operational parameters of the AGN-201 reactor, within regulatory limits but in
ways the DT had not previously seen, to simulate a potential bad actor nefariously
gaining access to the reactor to produce a dangerous isotope or other illicit use. The
DT analysts (blue team) then analyzed the resulting outputs from the reactor and
determined if/when anomalous actions were taken.
    The test consisted of three separate alterations to the reactor with different amounts
of reactivity change. From highest to lowest reactivity change, the list of alterations
to the reactor is:
. Inserting a polyethylene rod into the central irradiation facility.
. Inserting a cadmium foil cover into the central irradiation facility.
. Removing two graphite blocks from the thermal column.
   All the alterations were performed while the reactor was operating, and the
operator and reactor supervisor tried to make these changes as hidden as possible.
   After a sequence of first determining if the reactor was critical using a fine control
rod (FCR) and a coarse control rod (CCR), a “nefarious” was emulated by placing
cadmium foil cover into the central irradiation facility to act as a poison. The cadmium
was inserted into the reactor approximately 20 min after the polyethylene rod was
removed, resulting in the FCR.
The First Real-Time Digital Twin of a Nuclear Reactor                               245
To enhance the Digital Twin (DT) of the AGN-201 reactor with an effective anomaly
detection system, the team implemented the Isolation Forest machine learning algo-
rithm during the Red versus Blue Test. This algorithm is particularly good at spotting
unusual patterns in data without the need for previously identified examples of such
anomalies.
    The team first standardized the data, then fine-tuned the Isolation Forest model to
the specific characteristics of the DT’s data. They also set a ‘contamination parameter’
to estimate the expected ratio of anomalies in the data.
    As the reactor operates, the algorithm continuously analyzes incoming data,
deciding if each piece is normal or not. If it detects what seems to be an anomaly,
it marks and logs the occurrence for later review. This system is key in promptly
identifying and responding to any irregularities that might arise.
    The initial operation was used to assess the acceptability of the reactor physics
model and anomaly detection model for predicting off-normal conditions (i.e., simu-
lated proliferation attempts). Given the success of both models, two additional reactor
operations were performed to assess the online capabilities of the DT system. This
ensured that data was being transferred from the AGN-201 digital acquisition system
to DeepLynx, where it could be analyzed in near real-time. This process allowed the
reactor operators at ISU to operate for approximately an hour, and then within the
next 20 min, the data could be assessed and discussed between the two teams. The
ability to provide a near real-time assessment of reactor operations was a first for a
DT in nuclear reactor operations.
Overall, the initial red/blue team exercise was a success in showing that a Digital
Twin framework could effectively transfer data from AGN-201 via DeepLynx and
analyze it for off-normal operations. This was demonstrated for both offline and
online operations to provide confidence that the methodologies present would be
transferable to the near real-time detection of anomalies.
   Through the creation of a digital twin, the reactor physics surrogate and anomaly
detection models were able to capture the major anomalous behaviors that the reactor
operations team performed. The integrated anomaly detection algorithm within the
twin displayed a remarkable aptitude in capturing many of the off-normal opera-
tions, homing in on sensor fluctuations to identify divergences from the norm. This
achievement is particularly noteworthy given the complexity of the reactor’s oper-
ational dynamics and the nuanced variations that were successfully flagged, show-
casing the algorithm’s precise tuning and responsive nature to dynamic operational
changes, and the impact of using DT’s in this realm.
246                                                                             J. Darrington et al.
   Future work will entail wrapping in the PKE-SM to enhance the capabilities of
the reactor physics surrogate model. This will allow us to capture changes in power
as a function of time rather than just examining keff . For the reactor physics models,
an additional area of focus will be on determining what levels of perturbation are
quantifiable, meaning how small of a perturbation can be detected before it falls
within the range of uncertainty in the model, sensor data, etc. For both the reactor
physics surrogate and anomaly detection models, we plan to focus on detecting
startups, power changes, and shutdowns to prevent flagging these as anomalies.
   Through further testing, we were able to provide a near real-time assessment of
reactor operations and were able to detect nearly all the anomalous behaviors of the
reactor.
9.5 Conclusion
The Digital Twin of the Idaho State University’s AGN-201 was a resounding success,
despite the difficulties and hurdles in both operation in implementation. The Idaho
National Laboratory team was able to successfully create a near real-time twin of a
functioning nuclear reactor, a first in the industry. Not only did they create a way for
monitoring the twin using readily available cloud services but worked toward creating
a framework for monitoring the correct usage of said reactor to stop potential bad
actors.
   This work will lay the foundation for future digital twins in the Idaho National
Laboratory’s sphere of influence and has proven that they have the capability to
handle larger projects.
References
Abstract Over the past two decades, simulation and Digital Twin (DT) technolo-
gies have become increasingly prevalent in health care. These technologies have
significant potential to advance modern medicine, enhance clinical decision-making
and team performance, improve healthcare delivery, reduce cost, and improve
patient outcomes. This chapter provides an overview of these technologies and
their emerging applications in the field of health care, including opportunities to
accelerate discovery in basic science, delivery of more realistic training opportu-
nities that advance clinician competence and interprofessional teamwork, and an
efficient and cost-effective approach to analyze, improve, and monitor clinical work-
flows, healthcare delivery systems, and their performance. This emerging field fosters
multidisciplinary research among healthcare professionals, information technology
experts, engineers, and data scientists, all working together to better serve our society.
Widespread adoption of these technologies will require solutions to address several
technical, ethical, and regulatory challenges. These solutions will require a close
collaboration between industry, academic centers, and government to develop a
thoughtful approach that aligns implementation and integration of simulation and
DT technologies into the healthcare workplace with measurement of meaningful
outcomes to ensure broad-based access benefit from these important tools.
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2024          249
M. Grieves and E. Hua (eds.), Digital Twins, Simulation, and the Metaverse, Simulation
Foundations, Methods and Applications, https://doi.org/10.1007/978-3-031-69107-2_11
250                                                                       Y. Dong et al.
1 Introduction
Modern health care is a complex adaptive system that requires highly skilled inter-
professional team members to work together using advanced equipment and standard
processes to care for patients with a wide variety of diseases with steadily increasing
acuity over time. While the rapid pace of advancement in medical science offers
exciting possibilities and new hope to patients and their family within our global
community, it also demands constant changes in an already complicated and strained
healthcare delivery system. The practice of medicine and current healthcare outcomes
fall far below the level of quality and safety achieved by high-reliability industries,
such as aerospace and manufacturing. As a result, institutions have become increas-
ingly focused on the “value equation” of the care they deliver, generally defined as
the quality of care divided by the total cost of patient care over time [1].
    Over the past two decades, simulation and Digital Twin (DT) technology and
their applications have become increasingly prevalent in health care. These tools
can provide increasingly accurate representations of the pathobiology behind patient
conditions, opportunities to monitor patient health and optimize disease diagnosis and
management, enhance and evaluate the performance of clinicians and their healthcare
teams, and analyze and improve healthcare delivery and systems. Using a systematic
implementation strategy that includes key stakeholder collaboration, these technolo-
gies offer the opportunity to accelerate medical research discovery and improve
healthcare system performance and patient outcomes with greater efficiency and at
a reduced cost. This chapter provides an overview of these technologies and their
emerging applications in health care.
to scale. However, these healthcare improvements come with their own set of chal-
lenges. Population aging and growth, the increasing prevalence of chronic condi-
tions, and advanced diagnostic and treatment modalities are both driving cost and
reducing access to and equity of care. As the complexity of disease management
has increased, there is a growing demand for a collaborative, multidisciplinary, and
team-based approach that includes physicians, nurses, pharmacists, and respiratory
therapists and technicians. These teams require a clear understanding of each other’s
roles and responsibilities and a high level of communication and teamwork to ensure
timely and coordinated care delivery.
    Given its complexity, health care today has evolved into a “system of systems” of
healthcare delivery focusing on not only just safe and timely medical care, but also
better management of resources to meet demand at scale across geographic locations
and time. The Systems Engineering Initiative for Patient Safety (SEIPS) 3.0 frame-
work could serve as a guiding model in this complex landscape [3]. It offers a holistic
approach to integrating various components such as people, tasks, tools, and orga-
nization, thereby providing a comprehensive view of healthcare delivery systems.
These “systems” are also often stratified using different levels, including biology and
disease, patients and clinicians, and care delivery processes at the bedside, within
the hospital, and throughout healthcare systems as a whole [4]. A System of Systems
(SoS) in health care refers to integrating and interacting various independent systems
within the healthcare sector to achieve a more comprehensive, efficient, and effective
healthcare delivery.
    In complex healthcare systems, the overall system performance is directly linked
to the functionality of each subsystem and the interconnections between them. Indi-
vidual components, such as patient care, administrative processes, and healthcare
technology, must not only function effectively in isolation but also integrate seam-
lessly. The synergy between these subsystems is crucial for the overall system perfor-
mance. Medical errors and gaps in care quality have been well documented. However,
fixing them is not easy due to the complex and unpredictable work environment
common in medical practice [5, 6]. The recent global COVID-19 pandemic and its
impact further exposed these persistent vulnerabilities in healthcare systems world-
wide [7–9]. The pandemic has served as a catalyst for change, underscoring the
importance of improving current healthcare delivery mechanisms. It is imperative to
enhance the efficiency of healthcare delivery while maintaining the highest standards
of quality care and use innovative solutions and technologies to improve healthcare
access and distribution on a global scale. This systems view of healthcare delivery
also provides a framework that we can use to better understand how simulation and
DTs can be used to offer important solutions to these complex problems.
Over the past 30 years, we have witnessed a technological revolution that has signif-
icantly impacted all aspects of our lives, including health care. Personal computers,
252                                                                       Y. Dong et al.
DTs have various applications in health care, such as training, diagnosis, manage-
ment, and care delivery for patients, providers, and healthcare organizations. One
Evolution of Simulation and Digital Twin in Health Care: From …                      253
way to categorize DTs is based on Siemens’ framework, which has three levels of
twinning [15].
   First, product twining provides a virtual-physical connection to analyze how a
product performs under various conditions and adjust in the virtual world to ensure
that the physical product will perform exactly as planned in the field. For product twin-
ning, one example is the DTs of medical devices (e.g., digital radiological devices)
[16, 17]. These efforts focus on the physical device/product twinning (similar to the
concept of DTs for management in manufacturing) and are mostly used to monitor its
status, diagnose issues, and test solutions remotely, ultimately optimizing its perfor-
mance and reducing the risk of malfunctions. From digital devices to digital/virtual
patients, precision medicine is one area that exemplifies the application of both
product twinning and process twinning. Digitally replicating the human body (from
cell to organ biological/physiological systems) allows for in silico clinical trials to
examine the prevention, early detection, and targeted treatments of many diseases.
These DT human body/organ systems have been used for drug development and
treatment recommendations [18]. Another example of virtual patients is optimizing
health care at the individual level, i.e., “personalized health monitoring,” with the
goal of healthcare management and promoting healthy behavior. It can be used to
create personalized models of patients that can be used to monitor their health status,
predict the course of disease, and optimize treatment plans [19]. This application
is still focused on modeling a virtual patient instead of the healthcare system at
large and is mainly focused on healthy people’s daily life and their living environ-
ment instead of a specific healthcare facility. The third level is system twinning,
which includes using DTs to improve hospital operation processes and workflows
by allowing managers to tweak inputs and see how outputs are affected without the
risk of upending existing workflow [20]. The system performance can be captured,
analyzed, and acted on operational data, providing insights for informed decisions
to maintain effective interactions among the components of the system at the system
level.
DTs have risen at the intersection of Industry 4.0 and the Internet of Things (IoT),
relying on converging technologies in big data analytics, pervasive sensing, and cloud
computing infrastructure. In this new era, simulation and DTs promise to enhance
healthcare delivery by improving efficiency, flexibility, and patient-centered care, all
of which are crucial in a post-pandemic world. This book chapter highlights how
technological advancements have enriched health care with valuable data and digital
assets, making it imperative to utilize these resources to optimize the entire spectrum
of healthcare services. We will delve into the applications, benefits, challenges, and
future prospects of DTs and simulation technologies in health care, particularly in
the context of a post-COVID-19 world acute care setting. We will explore how these
technologies can serve as invaluable tools for healthcare professionals, clinicians,
254                                                                      Y. Dong et al.
Modeling and simulation have emerged as a powerful tool for understanding complex
biological systems and processes in addition to the traditional theory and experiment
[21]. Utilizing mathematical models, simulation allows researchers and clinicians
to predict disease outcomes, test treatment interventions, and optimize healthcare
delivery. There are two common approaches for physiology and disease modeling:
rule-based and data-driven modeling. Each of these approaches offers distinct advan-
tages and challenges [22]. The rule-based approach relies on predefined rules and
knowledge about biological systems to model disease processes. These rules are
typically derived from clinical guidelines, expert opinions, or well-established phys-
iological pathways and experienced-based learning in clinical settings. It has the
benefits of transparency (by incorporation of directed acyclic graphs in the creation
and execution of rules), easy control of model behavior, and stability for output,
which make it easy to accept for clinicians [23, 24]. At the same time, it also has
stability and flexibility limitations because it requires expert knowledge and comes
with a possibility of bias. The mechanistic model is designed to represent the under-
lying mechanisms of a system—the processes and interactions that lead to observed
phenomena. They are often contrasted with empirical models, which are purely based
on observed data without assumptions about the underlying physiology processes.
The data-driven model uses large datasets to investigate inter-relationships using
different statistical and data mining methods. It is based on systems’ biology and
systems’ immunology and has expanded rapidly in the last decade because of more
accessible data from EMR and more computer power [25, 26].
closely mimic real-world clinical scenarios based on physiology [11]. These plat-
forms are designed to be interactive, allowing medical students and healthcare profes-
sionals to engage in problem-solving, diagnosis, and treatment planning in a risk-free
environment. The virtual training modules often include a wide range of case studies
covering various diseases, conditions, and patient demographics, providing a compre-
hensive learning experience. The lab and vital signs will change based on the inter-
vention by user input. The virtual patient offers a practical understanding of disease
mechanisms and treatment protocols, thereby bridging the gap between academic
learning and clinical practice. This hands-on approach enhances the learner’s ability
to make informed clinical decisions in a safe environment that will not harm patients,
ultimately leading to better patient outcomes.
    Archimedes Diabetes Model was designed to simulate individual patients and
how they would respond to various treatments for diabetes. It incorporated a wide
range of variables, from glucose metabolism to treatment protocols, to generate
realistic patient outcomes [27, 28]. This model serves as one of the early successes
in utilizing computer-based physiology to understand and manage diabetes. This was
also an early implementation of using directed acyclic graphs to explain the complex
pathophysiology and rules based on the clinical environment. The model underwent
rigorous validation, and its predictions were found to be highly consistent with the
results of randomized controlled trials. This level of accuracy made the Archimedes
Diabetes Model a valuable tool for clinicians, researchers, and policymakers alike. It
not only helped in understanding the complex physiology of diabetes but also played
a crucial role in optimizing treatment plans and healthcare policies related to diabetes
management.
    Agent-based models can simulate the interactions between various biological
agents, such as cells and molecules, in each biological system [29, 30]. These agent-
based models are particularly useful for understanding complex, multiscale biolog-
ical systems where traditional modeling approaches may fall short. By providing a
more nuanced understanding of biological systems, agent-based models have paved
the way for more targeted and effective treatments, bridging the gap between basic
biological research and clinical applications. This approach is offering insights that
are directly applicable to patient care and treatment optimization. The research group
led by Vodovotz explores the use of omics data and mathematical modeling to under-
stand and explore various critical illness management strategies by integrating data-
driven and knowledge-based modeling approaches [22]. This approach aims to inte-
grate large-scale data with computational modeling to improve research and clinical
applications in diseases involving severe inflammation and immune responses.
    BioGears is an open-source, comprehensive human physiology engine that has
been instrumental in driving medical education, research, and training technolo-
gies [31]. BioGears aims to provide accurate and consistent physiology simulation
across the medical community. It can be used as a standalone application or inte-
grated with other simulators and sensor interfaces making it a valuable asset in the
healthcare simulation community. Similar to BioGears, the Pulse Physiology Engine
offers robust physics-based circuit and transport solvers [32]. It includes a common
data model for standard models and data definitions, a software interface for engine
256                                                                          Y. Dong et al.
some risk to the participants. In silico trials remove this concern entirely, as they
are conducted in a virtual environment. This not only ensures the safety of poten-
tial patients but also allows for the testing of treatments that might be considered
too risky for traditional trials. By offering a safer, faster, and more cost-effective
alternative to traditional methods, in silico clinical trials are poised to become a
cornerstone in the future of medical research. They offer a promising pathway for
the development of new treatments and medical technologies, with the potential to
significantly improve patient outcomes and healthcare systems globally. Important
components of biomedical ethics, such as informed consent, algorithm fairness and
biases, intellectual property law, data privacy, safety, and transparency, should be
considered alongside the regulatory issues of DTs [36]. Computer modeling and
simulation can be adopted from other industries to aid in various stages of medical
device development, testing, clinical evaluations, and failure analysis, leading to cost
reduction [37].
    Recently, the FDA has published reports to showcase how modeling and simula-
tion can be used for scientific research and regulatory decision-making [38]. FDA also
published guidance providing recommendations to assess the Credibility of Compu-
tational Modeling and Simulation in Medical Device Submissions [39]. Virtual clin-
ical trials represent a significant shift in how clinical research is conducted. By
leveraging DT platforms, researchers can gather large data more quickly and from a
broader population base. Simulations and virtual environments offer a unique oppor-
tunity for policymakers to test the implications of healthcare policies in a controlled,
risk-free setting [40]. By modeling the outcomes of proposed policies, stakeholders
can anticipate their effects and refine them before implementation, ensuring that new
regulations are both effective and efficient.
Over the past two decades, there has been a significant expansion in clinical training
for healthcare professionals through various simulation technologies. These tech-
nologies range from task trainers, mannequins to more advanced Virtual Reality
(VR) and Augmented Reality (AR) systems. The learners include the multidisci-
plinary clinical team (physicians, nurses, pharmacists, and respiratory therapists).
While traditional task trainers have been around for many years, modern technology
has allowed for the expansion and sophistication of simulation training programs.
These programs now cover a wide array of skills, from procedural skills to non-
technical skills like communication, compassion, and empathy. Numerous studies
have shown that simulation-based training is associated with improved care processes
and better patient outcomes [41].
258                                                                        Y. Dong et al.
Modern mannequins used in medical training have evolved to become highly sophis-
ticated, computer-driven devices. These mannequins can be controlled remotely or
in-room, offering a range of clinical scenarios for trainees to practice clinical skills
and teamwork [42, 43]. Those mannequins provide invaluable hands-on experience
for clinicians in a controlled, risk-free environment for various specialties and disci-
plines [41, 44, 45]. The COVID-19 pandemic has forced the healthcare industry to
adopt distance simulation using various technologies to find alternative approaches
for traditional onsite clinical training [46, 47]. Simulation has been used for skill
assessment and evaluation in many disciplines [48–50]. The use of VR and AR
in healthcare training is growing rapidly to offer immersive learning experiences
for learners that traditional methods cannot match [51–53]. Game-based training
modules have also gained popularity, providing an engaging way for healthcare
providers to hone their skills [54]. Although the cost of hardware can be a limiting
factor, the benefits, such as improved skill retention and real-world applicability,
often outweigh the initial investment.
Simulation technologies not only provide clinical skill training but also serve as
research tools for human factor analysis. They offer a reproducible clinical environ-
ment where the interaction of system factors (clinician, technology, and workflow)
can be investigated independently of patient factors [55, 56]. Researchers can use
these technologies to investigate system vulnerabilities, test new workflows, new
technology innovations, and even simulate the impact of potential policy changes
before clinical implementation [57–60]. This simulation-based research extends the
utility of simulation technologies as integral tools for overall healthcare system
improvement [61].
The past 15 years have seen a data revolution in health care fueled by electronic
medical records’ (EMRs) implementation and adoption. This shift from paper
to digital documentation has unlocked a wealth of patient data, improving team
communication and clinical decision support, informing disease treatment plans,
and tracking outcomes. Further, the integration of EMRs with the IoT and Radio
Evolution of Simulation and Digital Twin in Health Care: From …                     259
anomalies [68, 73]. Healthcare leaders can shift from reactive decisions to proactive
optimization based on data-driven insights from DTs. Hospital-level model by using
predictive decision support model that employs real-time service data is drawn from
the systems and devices [74]. Their model enables assessing the efficiency of existing
healthcare delivery systems and evaluating the impact of changes in services without
disrupting the daily activities of the hospital. Along the same line, Karakra et al. 2020
developed discrete event simulation and DTs through a system called HospiT’Win
that allows for tracking the pathways of patients inside the healthcare organization to
manage growing demand and decrease waiting times [69]. Rodriguez-Aguilar et al.
2020 proposed a digital healthcare system initiative through multi-paradigm simu-
lation [75]. Computer modeling has also been used to simulate infectious disease
transmission dynamics, optimize the vaccination strategy, and test public policies
before clinical implementation [76–78].
The adoption of simulation and DT technologies in health care is not without its chal-
lenges [79, 80]. One of the most pressing issues is the need for accurate and complete
data. Incomplete or erroneous data can significantly impact the effectiveness of these
technologies in both training and real-world applications. Also, modeling complex
biological systems involves numerous variables and nonlinear interactions, making
it challenging. Regulatory hurdles, such as compliance with healthcare standards
and data protection laws, further complicate the adoption process. Ethical consider-
ations are paramount when using patient data in the DTs, especially when it comes
to patient and clinician data privacy and confidentiality. The collection and use of
patient data must adhere to strict ethical guidelines to ensure that individual privacy
is respected. Moreover, there is a risk of bias in data collection and analysis, which
could inadvertently lead to unequal healthcare delivery. Addressing these ethical
and bias concerns is crucial for the responsible deployment of these technologies.
Data sharing and interoperability present another set of challenges. Different health-
care systems often use different data formats and standards, making integration a
complex task. The lack of interoperability can hinder the seamless exchange of
information, thereby limiting the effectiveness of simulation and DTs technologies
in a multi-system environment [81].
    For simulation and DT tools to be effectively integrated into health care, there
needs to be a set of clearly defined performance metrics for systems evaluation. These
metrics should measure impacts of simulation and DTs across various domains,
including disease outcomes, patient outcomes, system outcomes, and return on
investment. The development of such comprehensive evaluation metrics will enable
stakeholders to assess the effectiveness of DT technologies objectively.
Evolution of Simulation and Digital Twin in Health Care: From …                      261
    Despite these challenges, there are numerous opportunities for innovation. For
instance, multiscale simulation is a powerful tool for understanding complex biolog-
ical systems and healthcare processes. By simulating the behavior of individual
molecules, cells, and tissues, multiscale simulations can provide insights into how
diseases develop and how drugs work. Multiscale simulations are being used to
understand the molecular basis of diseases, such as Alzheimer’s disease and drug
development [82, 83]. Meanwhile, DTs can be used to track the progress of patients,
monitor their vital signs, and predict when they may need medical attention. DTs have
been used to monitor cardiac and cancer patients for personalized treatment planning
[84, 85]. Additionally, AI algorithms can be used to analyze the data generated by
simulations and DTs. This data can be used to train AI models to make predictions
and identify patterns. AI models can be used to improve the accuracy of diagnoses,
recommend treatments, and develop new drugs. The synergistic relationship between
multiscale simulation, DTs, and AI offers unprecedented capabilities in health care,
from real-time analytics to predictive and prescriptive modeling for disease preven-
tion, diagnosis, and management. The integration of multiscale simulation, DTs, and
AI is creating a new era of precision medicine and personalized health care. Data-
driven decision-making is revolutionizing disease management and patient care. By
leveraging predictive analytics, healthcare providers can anticipate disease progres-
sion and intervene earlier. This proactive approach improves patient outcomes and
reduces the burden on healthcare systems. Simulation and DTs will empower clin-
icians and operational managers to use real-time data to streamline workflows and
enhance the quality of care. By analyzing patterns in healthcare delivery, they can
identify bottlenecks and inefficiencies, leading to policies that align with the goals
of all stakeholders and provide safe, effective, efficient, and equitable patient care.
    The multidisciplinary nature of this field allows for the convergence of health-
care professionals, engineers, data scientists, IT experts, AI specialists, etc. Such
collaboration can lead to innovative solutions that address the existing challenges.
Moreover, as these technologies become more integrated into health care, there will
be a growing need for workforce training and upskilling to ensure that healthcare
professionals can effectively leverage these advanced tools.
    While technological advances in health care are promising, their successful adop-
tion hinges on robust change management, implementation science strategies, and
workforce upskilling. Crucially, comprehensive training is needed to equip health-
care professionals with the skills to leverage new digital capabilities effectively. This
includes technical competencies, data literacy, human–AI collaboration, cyberse-
curity, and ethics. Investing in change management, implementation science, and
strategic workforce development will be critical for healthcare organizations to
capitalize on transformative technological advances. Several major societies and
groups are active in related fields. The Society for Simulation in Healthcare (SSH)
is a global community of medical educators who use various technologies, such as
manikins and task trainers, to deliver educational interventions for skills and team-
work training [86]. The Winter Simulation Conference includes many engineers
working on modeling and simulation in various industries [87]. The Interagency
262                                                                             Y. Dong et al.
6 Conclusion
The emergence of simulation and DT technologies marks the beginning of a new era in
health care. These technologies are practical tools that can improve disease manage-
ment, care planning, and resource allocation. While there are undoubtedly chal-
lenges, such as technical complexities, ethical and regulatory hurdles, overcoming
these obstacles will enable us to fully integrate simulation and DT technologies to
support healthcare system digital transformation that serves all patients with safety,
effectiveness, patient-centeredness, timeliness, efficiency, and equity.
References
 1. Yong PL., Olsen L, Jm M (2010) Value in health care: accounting for cost, quality, safety,
    outcomes, and innovation: workshop summary
 2. Brandt AM (2012) A reader’s guide to 200 years of the New England journal of medicine. N
    Engl J Med 366(1):1–7
 3. Carayon P et al (2020) SEIPS 3.0: human-centered design of the patient journey for patient
    safety. Appl Ergon 84:103033
Evolution of Simulation and Digital Twin in Health Care: From …                                  263
 4. Wickramasinghe N et al (2007) Healthcare system of systems. IEEE Int Conf Syst Syst Eng
    2007:1–6
 5. Institute of Medicine Committee on Quality of Health Care (2000) In A., To Err is human:
    building a safer health system, In: Kohn LT, Corrigan JM, Donaldson MS (eds) To Err is Human:
    building a safer health system. National Academies Press, National Academy of Sciences,
    Washington (DC)
 6. Makary MA, Daniel M (2016) Medical error—the third leading cause of death in the US. Br
    Med J 353
 7. Blumenthal D et al (2020) Covid-19—Implications for the health care system. N Engl J Med
    383(15):1483–1488
 8. Kaye AD et al (2020) Economic impact of COVID-19 pandemic on healthcare facilities and
    systems: international perspectives. Best Pract Res Clin Anaesthesiol 35:293–306
 9. Domecq JP et al (2021) Outcomes of patients with coronavirus disease 2019 receiving organ
    support therapies: the international viral infection and respiratory illness universal study
    registry. Crit Care Med 49(3)
10. Society for simuation in healthacre: what is simulation? (2024) 22 Apirl 2013. Available from:
    http://ssih.org/about-simulation
11. Cook DA, Erwin PJ, Triola MM (2010) Computerized virtual patients in health professions
    education: a systematic review and meta-analysis. Acad Med 85(10):1589–1602
12. Digital Twin Consortium (2024) Available from: https://www.digitaltwinconsortium.org/faq/
13. Kritzinger W et al (2018) Digital Twin in manufacturing: a categorical literature review and
    classification. IFAC-PapersOnLine 51(11):1016–1022
14. Cockrell C et al (2022) Generating synthetic data with a mechanism-based critical illness digital
    twin: demonstration for post traumatic acute respiratory distress syndrome. bioRxiv, p 517524,
    22 Nov 2022
15. SIEMENS (2024) What is a digital twin? 2024; Available from: https://www.sw.siemens.com/
    en-US/technology/digital-twin/
16. Pesapane F et al (2022) Digital twins in radiology. J Clin Med 11(21)
17. Bjelland Ø et al (2022) Toward a digital twin for arthroscopic knee surgery: a systematic review.
    IEEE Access 10:45029–45052
18. Hernandez-Boussard T et al (2021) Digital twins for predictive oncology will be a paradigm
    shift for precision cancer care. Nat Med 27(12):2065–2066
19. Rivera LF et al (2019) Towards continuous monitoring in personalized healthcare through
    digital twins. In: Conference of the centre for advanced studies on collaborative research
20. Vallée A (2023) Digital twin for healthcare systems. Frontiers Digital Health 5
21. Pool R (1992) The third branch of science debuts. Science 256:44–47
22. Vodovotz Y (2023) Towards systems immunology of critical illness at scale: from single cell
    ‘omics to digital twins. Trends Immunol 44(5):345–355
23. Lal A et al (2020) Development and verification of a digital twin patient model to predict
    specific treatment response during the first 24 hours of sepsis. Crit Care Explor 2(11):e0249
24. Lal A et al (2020) Artificial intelligence and computer simulation models in critical illness.
    World J Crit Care Med 9(2):13–19
25. Alquraishi M, Sorger PK (2021) Differentiable biology: using deep learning for biophysics-
    based and data-driven modeling of molecular mechanisms. Nat Methods 18:1169–1180
26. Yue R, Dutta A (2022) Computational systems biology in disease modeling and control, review
    and perspectives. NPJ Syst Biol Appl 8
27. Eddy DM, Schlessinger L (2003) Validation of the archimedes diabetes model. Diabetes Care
    26(11):3102–3110
28. Schlessinger L, Eddy DM (2002) Archimedes: a new model for simulating health care systems–
    the mathematical formulation. J Biomed Inform 35(1):37–50
29. An G (2022) Specialty grand challenge: what it will take to cross the valley of death: transla-
    tional systems biology, “True” precision medicine, medical digital twins, artificial intelligence
    and in silico clinical trials. Frontiers Syst Biol 2
264                                                                                       Y. Dong et al.
30. An G et al (2009) Agent-based models in translational systems biology. Wiley Interdiscip Rev
    Syst Biol Med 1(2):159–171
31. McDaniel M et al (2019) A whole-body mathematical model of sepsis progression and treatment
    designed in the BioGears physiology engine. Front Physiol 10
32. Bray A et al (2019) Pulse physiology engine: an open-source software platform for computa-
    tional modeling of human medical simulation. SN Compr Clin Med 1:362–377
33. Rovati L et al (2024) Development and usability testing of a patient digital twin for critical
    care education: a mixed methods study. Front Med 10
34. Pappalardo F et al (2019) In silico clinical trials: concepts and early adoptions. Briefings Bioinf
35. Viceconti M et al (2021) In silico trials: verification, validation and uncertainty quantification of
    predictive models used in the regulatory evaluation of biomedical products. Methods 185:120–
    127
36. Lal A et al (2022) Regulatory oversight and ethical concerns surrounding software as medical
    device (SaMD) and digital twin technology in healthcare. Ann Transl Med 10(18):950
37. Landscape report and industry survey on the use of computational modeling and simulation in
    medical device development (2023). Available from: https://mdic.org/resource/cmslandscape
    report/#download_form
38. Successes and opportunities in modeling and simulation for FDA (2022)
39. Guidance document (2023) Assessing the credibility of computational modeling and simulation
    in medical device submissions, U.S.F.D. Administration, Editor
40. Romero-Brufau S et al (2021) Public health impact of delaying second dose of BNT162b2 or
    mRNA-1273 covid-19 vaccine: simulation agent based modeling study. BMJ 373:n1087
41. Cook DA et al (2011) Technology-enhanced simulation for health professions education: a
    systematic review and meta-analysis. JAMA 306(9):978–988
42. Cooper JB, Taqueti VR (2004) A brief history of the development of mannequin simulators
    for clinical education and training. Qual Saf Health Care 13:i11–i18
43. Gaba DM (2004) The future vision of simulation in health care. Qual Saf Health Care 13:i10–i12
44. Schmidt E et al (2013) Simulation exercises as a patient safety strategy a systematic review.
    Ann Intern Med 158(5_Part_2):426–432
45. McGaghie WC et al (2014) A critical review of simulation-based mastery learning with
    translational outcomes. Med Educ 48:375–385
46. Buléon C et al (2022) The state of distance healthcare simulation during the COVID-19
    pandemic: results of an international survey. Adv Simul 7(1):10
47. Flurin L et al (2022) International virtual simulation education in critical care during COVID-19
    pandemic: preliminary description of the virtual checklist for early recognition and treatment
    of acute illness and iNjury program. Simul Healthc 17(3):205–207
48. Ryall T, Judd BK, Gordon CJ (2016) Simulation-based assessments in health professional
    education: a systematic review. J Mult Healthc 9:69–82
49. Boulet JR, Murray DJ (2010) Simulation-based assessment in anesthesiology: requirements
    for practical implementation. Anesthesiology 112(4):1041–1052
50. Hayden JK et al (2014) The NCSBN national simulation study: a longitudinal, randomized,
    controlled study replacing clinical hours with simulation in Prelicensure nursing education. J
    Nurs Regul 5(2, Supplement):S3–S40
51. Kononowicz AA et al (2019) Virtual patient simulations in health professions education:
    systematic review and meta-analysis by the digital health education collaboration. J Med Int
    Res 21
52. Foronda CL et al (2024) A comparison of virtual reality to traditional simulation in health
    professions education: a systematic review. Simul Healthc 19(1s):S90-s97
53. Woodall WJ et al (2024) Does extended reality simulation improve surgical/procedural learning
    and patient outcomes when compared with standard training methods?: A systematic review.
    Simul Healthc J Soc Simul Healthc 19(1S):S98–S111
54. Gentry SV et al (2019) Serious gaming and gamification education in health professions:
    systematic review. J Med Internet Res 21(3):e12994
Evolution of Simulation and Digital Twin in Health Care: From …                                 265
55. Small SD (2007) Simulation applications for human factors and systems evaluation. Anesthe-
    siol Clin 25(2):237–259
56. Deutsch ES et al (2016) Leveraging health care simulation technology for human factors
    research: closing the gap between lab and bedside. Hum Factors 58(7):1082–1095
57. Ahmed A et al (2011) The effect of two different electronic health record user interfaces
    on intensive care provider task load, errors of cognition, and performance. Crit Care Med
    39(7):1626–1634
58. Arriaga AF et al (2013) Simulation-based trial of surgical-crisis checklists. N Engl J Med
    368(3):246–253
59. Patterson MD et al (2013) In situ simulation: detection of safety threats and teamwork training
    in a high risk emergency department. BMJ Qual Saf 22(6):468–477
60. Yeh VJ et al (2024) Using mobile and remote simulation as a research methodology for health
    care delivery research. Simul Healthc
61. Goldshtein D et al (2020) In situ simulation and its effects on patient outcomes: a systematic
    review. BMJ Simul Technol Enhanc Learn 6(1):3–9
62. Asamoah DA et al (2016) RFID-based information visibility for hospital operations: exploring
    its positive effects using discrete event simulation. Health Care Manag Sci 21:305–316
63. Haddara M, Staaby A (2018) RFID applications and adoptions in healthcare: a review on patient
    safety. Procedia Comput Sci 138:80–88
64. Abkari SE, Jilbab A, Mhamdi JE (2020) RFID medication management system in hospitals.
    Int J Online Biomed Eng 16:155–168
65. Sutton RT et al (2020) An overview of clinical decision support systems: benefits, risks, and
    strategies for success. NPJ Digital Med 3
66. Cristofoletti M et al (2023) Towards a computer vision-based approach for digital twin
    implementation. Int Conf Intell Metaverse Technol Appl (iMETA) 2023:1–6
67. Wang Y et al (2024) TWIN-GPT: digital twins for clinical trials via large language model
68. Elkefi S, Asan O (2022) Digital twins for managing health care systems: rapid literature review.
    J Med Internet Res 24(8):e37641
69. Karakra A et al (2019) HospiT’Win: a predictive simulation-based digital twin for patients
    pathways in hospital. IEEE EMBS Int Conf Biomed Health Informatics (BHI) 2019:1–4
70. Peng Y et al (2020) Digital twin hospital buildings: an exemplary case study through continuous
    lifecycle integration. Adv Civil Eng 2020:1–13
71. Croatti A et al (2020) On the integration of agents and digital twins in healthcare. J Med Syst
    44(9):161
72. Augusto V, Murgier M, Viallon A (2018) A modelling and simulation framework for intelligent
    control of emergency units in the case of major crisis. Winter Simul Conf (WSC) 2018:2495–
    2506
73. Zhong X et al (2022) A multidisciplinary approach to the development of digital twin models
    of critical care delivery in intensive care units. Int J Prod Res 60(13):4197–4213
74. Karakra A et al (2018) Pervasive computing integrated discrete event simulation for a hospital
    digital twin. In: 2018 IEEE/ACS 15th international conference on computer systems and
    applications (AICCSA), p 1–6
75. Rodríguez-Aguilar R, Marmolejo-Saucedo JA (2020) Conceptual framework of digital health
    public emergency system: digital twins and multiparadigm simulation. EAI Endorsed Trans
    Pervasive Health Technol 6:e3
76. Currie CSM et al (2020) How simulation modelling can help reduce the impact of COVID-19.
    J Simul 14:83–97
77. Romero-Brufau S et al (2021) Public health impact of delaying second dose of BNT162b2 or
    mRNA-1273 covid-19 vaccine: simulation agent based modeling study. The BMJ 373
78. Zhang T et al (2023) Data-driven modeling and analysis for COVID-19 pandemic hospital beds
    planning. IEEE Trans Autom Sci Eng 20:1551–1564
79. Fackler J, Hankin J, Young T (2012) Why healthcare professionals are slow to adopt modeling
    and simulation. In: Proceedings of the 2012 winter simulation conference (WSC)
266                                                                                 Y. Dong et al.
80. Combs CD, Combs PF (2019) Emerging roles of virtual patients in the age of AI. AMA J
    Ethics 21(2):E153-159
81. Jacoby M, Usländer T (2020) Digital twin and internet of things—Current standards landscape.
    Appl Sci
82. Stefanovski L et al (2021) Scales in Alzheimer’s disease: biological framework for brain
    simulation with the virtual brain. Frontiers Neuroinformatics 15
83. Joshi A et al (2020) Opportunities for multiscale computational modelling of serotonergic drug
    effects in Alzheimer’s disease. Neuropharmacology 174:108118
84. Coorey G et al (2022) The health digital twin to tackle cardiovascular disease-a review of an
    emerging interdisciplinary field. NPJ Digit Med 5(1):126
85. Stahlberg EA et al (2022) Exploring approaches for predictive cancer patient digital twins:
    Opportunities for collaboration and innovation. Front Digit Health 4:1007784
86. Society for Simulation in Healthcare (SSH) Council for accreditation of healthcare simulation
    programs accreditation standards
87. 2024 Winter Simulation Conference (2024). Available from: https://meetings.informs.org/wor
    dpress/wsc2024/
88. IMAG: Interagency Modeling and Analysis Group (2024)
89. Virtual Physiological Human Institute (2024). Available from: https://www.vph-institute.org/
90. Europe Digital Twin in Healthcare (2024). Available from: https://www.digitaleurope.org/pro
    jects/ecosystem-digital-twins-in-healthcare-edith/
91. National Academies of Sciences, Engineering, and Medicine. Foundational Research Gaps and
    Future Directions for Digital Twins (2024). Available from: https://www.nationalacademies.
    org/our-work/foundational-research-gaps-and-future-directions-for-digital-twins
Digital Twins for Sustainable
Semiconductor Manufacturing
Ala Moradian
Abstract This chapter explores the applications of Digital Twins in smart semi-
conductor manufacturing, highlighting their potential to drive a more sustain-
able industry. Digital Twins enable advanced monitoring of chemical and energy
consumption, as well as other environmental impacts within semiconductor fabrica-
tion processes. The chapter reviews various initiatives and applications that utilize
Digital Twins for smart monitoring and provides guidance on extending these efforts
to further optimize semiconductor manufacturing. This includes enhancing device
performance, yield, and cost of ownership, while minimizing environmental impact.
The chapter concludes that achieving truly smart manufacturing requires a holistic
approach focused on optimizing specific equipment design and processes (such as
deposition, etch, lithography, chemical mechanical planarization) and fostering cohe-
sive collaboration among stakeholders from academia and industry. Addressing the
gaps in understanding physics, chemistry, sensor technology, software infrastructure,
data security, and establishing universal standards and protocols for data sharing
and integration are essential for realizing the full potential of Digital Twins in the
semiconductor industry.
1 Introduction
With the arrival of digital transformation innovations such as, Generative AI1 and
LLMs,2 Industry 4.0, autonomous vehicles, and Digital Twins the demand for semi-
conductor chip manufacturing continues to grow. As a result, sustainability measures
1   Artificial Intelligence.
2   Large Language Models.
A. Moradian (B)
Applied Materials Inc, Santa Clara, USA
e-mail: [email protected]
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2024          267
M. Grieves and E. Hua (eds.), Digital Twins, Simulation, and the Metaverse, Simulation
Foundations, Methods and Applications, https://doi.org/10.1007/978-3-031-69107-2_12
268                                                                       A. Moradian
and reducing the environmental impact will become a key area of focus for manu-
facturers as well as governments and local energy regulators. To address this sustain-
ability challenge, manufacturers along the entire value chain need to work both indi-
vidually as organizations and collectively as an industry to define goals and agree on
actions. Major semiconductor manufacturers have begun advocating for the industry
to enable a smart and digital future while simultaneously reducing environmental
impacts.
   In situ monitoring of energy, water and gas consumption of semiconductor manu-
facturing plants will be the first critical step. As the famous Lord Kelvin once
mentioned, what you cannot measure you cannot improve. By synchronizing data
from multiple sensors, and leveraging domain knowledge and models, detailed on-
demand reports of carbon footprint and related consumption parameters from both
the fab and sub-fab can be generated. In addition to the monitoring the consumptions,
Digital Twins of manufacturing equipment can enable what-if scenarios with virtual
wafer processing and estimating their environmental impact. Ultimately, manufac-
turing equipment in the fab and the auxiliary equipment in the subfab would need to
be enhanced with built-in intelligence to optimize process and operational conditions
for not only on-wafer process outcome but also the environmental impact of the whole
manufacturing flow of future semiconductor devices. This sustainability-driven goal
would require calibrated processes and hardware Digital Twins that can realistically
predict impact of any changes to the system, not only at the environmental impact
level, but also for the integration and performance of the semiconductor devices.
3   Internet of Things.
Digital Twins for Sustainable Semiconductor Manufacturing                            269
The concept of Digital Twins was first proposed by Michael Grieves in 2002 in
the context of product lifecycle management (PLM) [1]. Since then, there has been
multiple definitions suggested among which the definition by NASA and Grieves
are more broadly accepted. NASA described the Digital Twin as follows: “A Digital
Twin is an integrated multiphysics, multiscale, probabilistic simulation of an as-built
vehicle or system that uses the best available physical models, sensor updates, fleet
history, etc., to mirror the life of its corresponding flying twin” [2]. Grieves proposed
that the basic Digital Twin model has three components, viz. physical entity in real
space, virtual entity in virtual space, and the networks of information data that tie
the physical and virtual entities or space together [3]. Digital Twin (DT) constructs a
virtual replica of a physical scenario to monitor the behavior of the ongoing process,
with condition monitoring, detection of anomalies, and prediction of future trends
within the process. Based on the literature, the definitions of the Digital Twin is (see
Fig. 1):
. the digital representation of a physical entity with possible data and models. The
  term “data” refers to the data from all the processes obtained during the run time
  and the system’s development phase [4];
. it always incorporates synchronization with its related physical entities [5];
. it is a simulation model for the working of the related physical entity or process
  [6].
   The Digital Twin Consortium (DTC) which promotes awareness, adoption, inter-
operability, and development of Digital Twin technology, defined Digital Twin as a
virtual representation of real-world entities and processes, synchronized at a specified
frequency and fidelity [7].
    Another definition is based on ISO 23247 [8], which described a Digital Twin
as a “fit-for-purpose digital representation of an observable manufacturing element
(OME) with synchronization between the OME and its digital representation”. OMEs
include personnel, equipment, materials, processes, facilities, environment, prod-
ucts. Figure 2 shows a diagram of Digital Twin in manufacturing based on this
definition. According to a report by National Academy of Sciences focused on the
foundational research gaps and future directions for Digital Twins, the definition
was proposed as [14]: a set of virtual information constructs that mimics the struc-
ture, context, and behavior of a natural, engineered, or social system (or system-of-
systems), is dynamically updated with data from its physical twin, has a predictive
capability, and informs decisions that realize value. The bidirectional interaction
between the virtual and the physical is central to the digital twin.
    Feng et al. 2023 demonstrated an implementation of the ISO23247 for data
requirements for Digital Twins in additive manufacturing [8].
    DTC introduced the Digital Twin Capabilities Periodic Table (CPT) which is an
architecture and technology agnostic requirements definition framework, see Fig. 3.
It is aimed at organizations who want to design, develop, deploy, and operate Digital
Twins based on use case capability requirements versus the features of technology
solutions [9].
    Several systematic bibliometric analyses have been conducted in the litera-
ture targeting Digital Twin framework in the context of smart manufacturing. As
an example, Warke et al. 2021 surveyed evolution, background, and implementa-
tion of critical enabling technologies such as data-driven decision-making, machine
learning, artificial intelligence, and deep learning. In their study, Warke et al. 2021
leveraged Scopus and Web of Science databases from 2016–21 for their analysis [10].
Dalibor et al. 2022 presented a systematic mapping of 356 Digital Twins publications
to characterize the applications in different domains. 20 different application domains
were presented, but only 70% of the studied work corresponded to the manufacturing
industry. Those applications were related to monitoring and controlling either before
production, to improve the design process, or during product lifetime [11].
subject matter experts (SMEs) in their digitization efforts to deploy Digital Twins.
They are a consortium of 18 partners from all over Europe and promote the idea
that the concept of Digital Twin is one of the big game-changers in manufacturing
allowing companies to significantly increase their global competitiveness [12].
   With a value of $12.9 billion in 2022, the Digital Twin market is expected to
grow at a CAGR of 35% to 40% through 2030, according to industry analysts. This
market growth is primarily driven by the increasing adoption of enabling technologies
such as AI, enterprise internet of things (IoT) platforms, augmented reality (AR),
and virtual reality (VR) [13]. In a comprehensive report by National Academies
(funded by the Department of Defense—Air Force Office of Scientific Research
and Defense Advanced Research Projects Agency, the Department of Energy, the
National Institutes of Health, and the National Science Foundation) the opportunities
enabled by Digital Twin were defined and the foundational research and resources
needed to support the development of Digital Twin technologies were identified
[14]. The report presented critical future research priorities and an interdisciplinary
research agenda for the field, including how federal agencies and researchers across
domains can best collaborate.
Digital Twins play a pivotal role in enhancing sustainability efforts across various
industries. The ability to simulate and analyze operations in real-time allows busi-
nesses to identify and implement sustainable practices. As industries increasingly
prioritize environmental responsibility, Digital Twins emerge as powerful tools to
monitor, analyze, and minimize carbon footprints across diverse sectors. In manufac-
turing, these twins optimize energy consumption and production efficiency, reducing
carbon footprints. In agriculture, Digital Twins aid precision farming, minimizing
resource wastage. Smart buildings employ Digital Twins to optimize energy usage,
contributing to eco-friendly urban development. Below is a few examples.
   Data from two enterprises was used as a case study in an industrial DevOps
research project called Titan [15], which was focused on methods and tools for inte-
grating and analyzing big data from IoT devices in industrial manufacturing and
used to estimate power consumption. The authors proposed measures that can be
implemented in an industrial DevOps analytics platform, the Titan Control Center
[16]. Henning et al. 2021 discussed two industrial pilot cases, where analyzing power
consumption data can serve the goals of reporting, optimization, fault detection, and
predictive maintenance. In a pilot implementation of a power consumption analytics
platform, they showed how measuring real-time data processing, multilevel moni-
toring, temporal aggregation, correlation, anomaly detection, forecasting, visual-
ization can be implemented with a microservice-based architecture (see Figure 5),
stream processing techniques, and the fog computing paradigm, see Fig. 6 [16].
Digital Twins for Sustainable Semiconductor Manufacturing                                    273
Fig. 5 Microservice-based pilot architecture of the Titan Control center for analyzing electrical
power consumption [16]
   Seegrun et al. 2023 conducted a systematic literature review (SLR) through Scopus
and Web of Science. Their query retrieved 523 records on Scopus and 281 on Web of
Science (data extraction in October 2022). After the elimination of 249 duplicates,
the dataset contained 555 records. These publications were categorized according to
their industry focus.
   From the search results, application of Digital Twins were allocated throughout a
product’s lifecycle and assessed in terms of their technological maturity and sustain-
ability scope. To analyze the state of the art of research on Digital Twins in the context
of sustainability and to frame directions for future research, a SLR was conducted
to address the following research questions (RQs):
. RQ1: What are the application scenarios of Digital Twins in the context of
  sustainability in the manufacturing industry?
. RQ2: In which product lifecycle (PLC) phases are Digital Twins applied in the
  context of sustainability?
. RQ3: What is the technology readiness level (TRL) of the Digital Twin
  applications?
. RQ4: For the attainment of which sustainability objectives are Digital Twins
  currently applied?
. RQ5: What are the key directions for future research in the manufacturing industry
  for sustainability-focused applications of Digital Twins?
   The authors divided a product lifecycle into three stages: Beginning-of-life (BoL),
Middle-of-life (MoL) as well as End-of-Life (EoL) which are in turn subdivided into
the following lifecycle phases:
. BoL: Planning (BoL1), product development (BoL2), production (BoL3)
274                                                                        A. Moradian
Fig. 7 Distribution of
literature body according to
product lifecycle and
technology maturity level
[17]
Fig. 8 Application
potentials of Digital Twins
within the literature body
based on the stage in product
lifecycles [17]
    XMPro inc provides a generic Digital Twin platform that can track real-time data,
provide decision support From multiple data sources [18]. In one example, the plat-
form was used to provide a virtual representation of the wind farm (see Figure 9)
synchronized at a high frequency and fidelity, which enables real -time monitoring,
analysis, and optimization. By integrating data from sensors and historical perfor-
mance, the Digital Twin can predict the wind farm’s behavior and identify potential
issues before they occur, enabling proactive measures to be taken. The Digital Twin
remote operations center provides a platform for remote decision-making, allowing
operators to monitor and predict changes without disrupting operations, resulting in
improved performance, reduced maintenance costs, and enhanced safety.
    Abdune et al. 2023 proposed a data-driven methodology for integrating the energy
consumption model into Digital Twins using techniques such as segmentation and
regression. It relies on power absorption measurement of industrial equipment to
generate energy consumption related parameters to be fed into the DT model to
monitor the current operating condition of the physical system. A case study on an
industrial robot was used to validate and assess the performance of the approach in
a laboratory environment [20].
276                                                                      A. Moradian
Physics-based category
Data-Based Category
Data-based models are also referred to as data models, empirical models, or artifi-
cial intelligence methods. This category of Digital Twins mostly depend on time-
series statistical analyses and machine learning techniques. It can be considered as
Digital Twins for Sustainable Semiconductor Manufacturing                            277
proposes a heuristic approach for cognitive Digital Twin technology. The author
proposed the heuristic approach as a feature selection tool to enhance the cognitive
capabilities of a Digital Twin throughout the product design phase of production. The
proposed approach was validated using the use-case of Power Transfer Unit (PTU)
production [33].
Hybrid-Based Category
Hybrid modeling refers to the situation where part of a model can be formulated on the
basis of first principles and part of the model has to be inferred from data because of
a lack of understanding of the mechanistic details [34]. Huang et al. 2022 proposed
a modeling framework for hybrid learning-based Digital Twin for manufacturing
process with the corresponding trial implementation. The authors constructed a data
processing procedure to contextualize metadata sources across the process chain,
and a modeling pipeline for the integration of production domain knowledge and
AI techniques [35]. Langlotz et al. 2022 published a concept for Digital Twins that
are modeled by physics-based and data-driven models. Furthermore, this concept is
validated with the help of a use case that controls an energy management system of
a model-scale factory by finding the optimal time for charging and using the battery
[36].
    Reduced order models (ROM) or surrogate models are critical to enable real-time
Digital Twins and expedite the run-time of models. These methods are in partic-
ular critical for high-fidelity physics or hybrid models where solving massive set of
equations resulted from discretization of computational domain into thousands or
millions of grid point or mesh cells. Some of such methods are reviewed by Es-haghi
et al. 2024 [37].
    Another important area of research for enabling predictive Digital Twins, in
particular for applications where the data is scarce (or sparse) are AI-assisted
physics models, sometimes referred to as physics informed machine learning (PIML),
physics-informed neural networks (PINN). Raissi et al. 2019 are among the pioneers
of the PINN method [38]. In addition, Graph Neural Net (GNN) in another technique
with increased popularity. Based on inherent advantages of graph-based structures
for representing unstructured data, graph neural network has been receiving more and
more attention in predicting flow fields defined on unstructured computing nodes.
Some of these methods are based on Scarselli et al. 2009 [39] and Sanchez-Gonzalez
et al. 2020 [40].
State-Based Category
State-based models are related to the segmentation of the system into different func-
tions and sections, each of which is typically represented by an average behavior
Digital Twins for Sustainable Semiconductor Manufacturing                          279
The interTwin project is funded by the European Union Horizon Europe Program
(2023-25) aimed at designing and building a prototype of an interdisciplinary
Digital Twin engine (DTE), based on a co-designed Blueprint Architecture. The
mission of the Digital Twin engine is to simplify & accelerate the development of
complex application-specific Digital Twins that benefits researchers, business, and
civil society. The effort attempts to extend the technical capabilities of the European
Open Science Cloud with modeling & simulation tools, demonstrate data fusion with
complex modeling & prediction technologies among other things. The program has
described several deliverables, one of which is to support the climate change use
cases and the implementation of the related impact decision support tools. Some of
the use cases of the program include studying tropical storms change in response to
climate change—The goal is to provide notebooks for scientists and policy makers
for running analysis on tropical cyclones on future projection data. Other example
of use cases are wildfire risk assessment in response to climate change, and flood
early warning in coastal and inland regions [44].
   Another European research project called Trustworthy virtual experiments and
Digital Twins (ViDiT) was started, funded by the European Union through the Euro-
pean Partnership on Metrology, co-financed from the EU’s Horizon Europe Research
and Innovation Program and by the Participating States.
   ViDiT focuses on virtual experiments and Digital Twins as key enabling tech-
nologies to achieve and realize European strategic policies devoted to sustainability
and digitalization within the complex framework of Industry 4.0 and the European
Green Deal. The consortium of this three-year project consists of 21 institutions: 8
National metrology institutes, 2 research centers close to industry, 5 universities and
6 companies. The project is coordinated by Physikalisch-Technische Bundesanstalt
(PTB), based in Braunschweig, Germany [45].
   Green material optimal selection (GMOS) in product design is a key issue for
realizing sustainable manufacturing. To improve the accuracy and efficiency for
280                                                                               A. Moradian
Fig. 10 Optimal selection of green material for manufacturing based on Digital Twins [46]
green material optimal selection in product, a new method driven by Digital Twin
was proposed by Xiang et al. 2019 [46]. In their method, first a high-fidelity Digital
Twin of the physical product was developed. Then the actual data (physical) and
the cyber (modeling) data was combined, and the performance of the green material
selection was evaluated. The team used a laptop design example to showcase the
methodology, see Fig. 10.
   Tao et al. 2018 presented a five-dimensional model for Digital Twins. In Tao’s
model M_DT = (PE, VE, DD, Ss, CN), where PE refers to physical entity, VE
is virtual entity, DD is Digital Twin data, Ss is service for PE and VE, and CN is
connection among all PE, VE, and Ss. Xiang et al. 2019 proposed a variation of this
methods for GMOS where the 5D variation for Digital Twin model (5D-EDTM) is
divided into two parts M = (PE, VE, SE) + (DE, CE); one part consists of PE which
refers to evolutionary model for physical space, VE refers to evolutionary model for
virtual space, SE is the evolutionary model in service, DE stands for data evolutionary
model, CE refers to connection model among all dimensions. Therefore, 5D-EDTM
for GMOS can be divided into five models [6]. Implementation of the 5D-EDTM
driven GMOS method applied to GMOS is depicted in Fig. 11.
   As described earlier, Henning et al. [16] used the Titan architecture to monitor
and analyze the power consumption for two pilot cases. Davila et al. 2023 performed
a similar analysis included other types of energy different from electricity (i.e., gas,
heat, compressed air) and relevant resources for sustainability, such as water and
raw materials [47]. Their study also considered the privacy concerns of the data from
different sources or stakeholders. Basically, they proposed a pre-processing data to
Digital Twins for Sustainable Semiconductor Manufacturing                              281
Fig. 11 Digital Twin-based method for green material selection methodology, 5D-EDTM [46]
282                                                                           A. Moradian
obfuscate the protected information about the production process or the machines
used. There are many approaches to achieve privacy-preserving data sharing, for
example encryption, anonymization, perturbation, encoding or more complex data
synthesis (creating data synthetically). Choosing the appropriate technique depends
mainly on how the data is going to be used downstream, and what level of privacy,
or disclosure risk is desired.
    The methodology proposed by Davila et al. 2023 consists of three levels for the
development of the Digital Twin. In first level (connection and communication), a
process model of the manufacturing steps is created. The first step is defining the
system boundaries, starting with the raw materials, and ending with the finished
product. Afterward, all the required inputs and outputs of each of the machines
involved in the different production processes are determined. The different inputs
can be electricity, heat, gas, compressed air, or materials, within others. The output are
typically the processed materials, intermediate products and finally the product. All
the relevant measurement points within the process are determined, to install sensors
required for the measurement of the inputs. As a result of the process modeling,
the flow of resources into and out of the system boundaries can be measured and
monitored in real-time. The second level is concerned with data collection from
all the machines and different data sources (and types), such as sensors, machines,
production, supplier, and weather data. Finally, in the third level (i.e., cognitive, and
analytical level), the data is used to perform energy efficiency and sustainability
assessments and the costs and estimate the durations per product, see Fig. 12.
    In an example, Davila et al. 2023 demonstrated how different processes that
involve a set of machines or production steps can be considered for an overall energy
efficiency and sustainability analysis.
    Ma et al. 2022 proposed a sustainable smart manufacturing strategy based on
information management systems for energy-intensive industries (EIIs) from the
product lifecycle perspective. Their study encompasses a Digital Twin-driven opera-
tion mechanism, and an overall framework of big data cleansing and integration are
designed to describe sustainable smart manufacturing. By two examples, the authors
showed that the unit energy consumption and energy cost of production as well as the
costs of environmental protection can be decreased by implementing their proposed
strategy based on the ‘cradle-to-gate’ lifecycle data analysis [48].
    Li et al. 2022 developed a data-driven hybrid petri-net (DDHPN) [49, 95] inspired
by both the state-based energy modeling and machine learning for establishing
the energy behavior meta-model. Gaussian kernel extreme learning machine was
proposed to fit the instantaneous firing speed of energy consumption continuous
transitions in DDHPN. DDHPN-based energy behavior model was driven by phys-
ical data under real-time working conditions, operating parameters, and production
load for generating a virtual data space of energy management, See Figs. 13, 14, 15,
and 16.
    In the context of sustainable manufacturing, the Triple Bottom Line TBL supports
the concept of sustainability in industries by clarifying three dimensions: economic,
environmental, and social [50]. The concept of triple bottom line has been introduced
over a decade ago and is being discussed more recently in the context of Digital Twins
Digital Twins for Sustainable Semiconductor Manufacturing                              283
Fig. 12 Sustainability Digital Twin architecture proposed by Davila et al. 2023 [47]
for sustainability. For example, Miehe et al. 2021 attempted to stimulate interdisci-
plinary scientific discourse by discussing the interpretation of sustainable production
and the role of Digital Twins [51]. They concluded that although sustainability repre-
sents a key factor of future production, it is not conclusively defined in order to be
technically applicable. Existing (bottom-up) approaches assessing the contribution
of DTs to sustainable production are therefore not considered comprehensively from
a contemporary perspective. In order to navigate the technical development of DTs in
an appropriate direction in the sense of sustainable production, a further (top-down)
perspective is necessary.
   Chavez et al. 2022 identified suitable key performance indicators (KPIs) for a
Discrete Event Simulation (DES) model and evaluated the impact in a drone factory
in four scenarios that test final assembly processes. Based on this project (called
TWINGOALS and funded by European Institute of Innovation and Technology (EIT)
[5294]), the team integrated sustainability indicators in a simulation model of a drone
assembly cell at the Stena Industry Innovation Lab (SII-Lab), based at Chalmers
University of Technology in Gothenburg, Sweden. This study attempted to answer
284                                                                                A. Moradian
Fig. 13 Energy behavior modeling based on a data-driven hybrid petri-net by Li et al. 2022 [49]
Fig. 14 Energy management framework proposed by Li et al. 2022 for manufacturing [49]
Digital Twins for Sustainable Semiconductor Manufacturing                 285
Fig. 17 Incorporation of KPIs related to the triple bottom line principle in the product lifecycle
[53]
two questions first identifying the most suitable KPIs to assess the sustainability
of production systems through a DES (or Digital Twin), and second, identifying
improvement sustainability opportunities of production systems. See Fig. 17 for
an example of the selected KPIs to be used in a Digital Twin model with TBL
considerations.
   In a study performed by Banerji et al. 2021, a real-time agent-based optimization
motion planning for energy consumption of a robotic cellular was investigated. The
research proposed a framework a qualitative analysis, and a quantitative comparison
based on energy consumption in a robotic experiment case [54].
Since the first publication by Michael Grieves in 2005 [1] there has been remarkable
interests in developing and deploying Digital Twins to enable and accelerate smart
manufacturing across many industries. Similarly, there has been extensive progress
and developments by academia [55]. It is not a surprise that there has been an ever-
growing excitement also by the semiconductor industry to partake in adopting the
technology; however, semiconductor manufacturing is still in the very early innings
of creating and capturing value from Digital Twins with proven efficacy.
   In 2022, the US government implemented an industrial strategy to revitalize
domestic manufacturing, strengthen American supply chains, and accelerate the
industries of the future. The CHIPS and Science Act of 2022 [56], which buildsd on
this progress, making historic investments that would potentially poise U.S. workers,
local communities, and American businesses to win the race for the twenty-first
century. This act promises to strengthen American manufacturing, supply chains, and
national security, and invest in research and development, science, and technology
in the industries of tomorrow, including nanotechnology, clean energy, quantum
computing, and artificial intelligence.
Digital Twins for Sustainable Semiconductor Manufacturing                         287
Fig. 19 Classifications and characteristics of Digital Twins for semiconductor manufacturing [60]
million additional skilled workers will be needed to meet demand in the semicon-
ductor industry [65]. To capture the domain knowledge (i.e., codifying the know-
how) and sustainable workforce development, EduTwin™ and SemiGuru™ repre-
sent all the educational aspects of the twins, for example virtual unit or integrated
processing for training new process engineers by providing an accurate replica of a
specific equipment or process where a virtual wafer can be processed, and results
can be analyzed. The virtual-processing capability developed based on the chamber,
processes, or fleet of twins would support democratization of the semiconductor tech-
nology knowledge. In multiple scopes, focused on different segmentsof the work-
force (,e.g., STEM students in high-schools, community colleges, universities, or
talent onboarding), or various levels (e.g., introductory to advanced), EduTwin and
SemiGuru can prepare the much-needed workforce and “bend the learning curve”
in training and onboarding technicians and engineers. Educational platforms based
on Digital Twins can prepare the much-needed workforce and "bend the curve"
in training and onboarding technicians and engineers. To properly safeguard intel-
lectual property and proprietary know-how, organizations should develop both an
internal and an external versions of such training platforms. For instance, EduTwin
could function as a more generic educational platform for training new industry
entrants, while SemiGuru could serve as an advanced educational platform containing
valuable details and know-how through twins with higher fidelity, accuracy, and
focus. Advancing digital twin-enabled curricula for training a domestic semicon-
ductor workforce is one of the objectives envisioned for the CHIPS Digital Twins
Manufacturing USA Institute.
Digital Twins and virtual fabrication clearly offer great promise, but there are chal-
lenges that need to be overcome to tap their full potential in the semiconductor
manufacturing environment. Semiconductor equipment and processes are highly
multiphysics and multiscale environments, with a diverse set of length and times
scales. It is extremely challenging to integrate these disparate length- and time-
scale models. In many cases, a complete understanding of the underlying physics
and chemistry for processes used in the semiconductor manufacturing is lacking.
Therefore, a hybrid-model, that not only rely on the physics but also incorporate
avaiable test data, are more desired. Another fundamental hurdle is having adequate
sensors providing real-time data that can be placed within the harsh environments
of the process chambers. The absence of appropriate sensors makes it difficult to
build accurate predictive models and algorithms that account for all the nuanced
interdependencies and nonlinear effects. Finally, there is the challenge of gathering
adequate amount of useful data to train data models and capturing adequate level of
physics to matching the models accurately with actual on-wafer parameters. This is
exactly why Digital Twins are built based on the best models and will need to be
revised and improved whenever more accurate data becomes available or there is a
major change in our fundamental understanding of the processes.
Digital Twins for Sustainable Semiconductor Manufacturing                            291
Fig. 20 Schematic of Digital Twin for fab (FleetTwin) for heater jacker health and proper
installation [66]
Researchers led by Udit Gupta of Harvard University used publicly available sustain-
ability reports from companies including TSMC, Intel Corp. and Apple Inc. to
show that as computing becomes increasingly ubiquitous, “so does its environmental
impact.”
   Information and computing technology is expected to account for as much as
20% of global energy demand by 2030, with hardware responsible for more of
that footprint than the operation of a system. “Chip manufacturing, as opposed to
hardware use and energy consumption, accounts for most of the carbon output,”.
   Power use is more dramatic; TSMC’s annual electricity consumption is estimated
by Greenpeace at 4.8% of Taiwan’s entire usage, and more than that of the capital,
Taipei. Greenpeace says that will rise to 7.2% once commercial production comes
online of TSMC’s newest fabs that will shrink the process further from the current
leading-edge of 5 nm, or billions of a meter, to 3 nm chips [69].
   It is also expected to consume 237 TWh of electricity globally in 2030, close to
Australia’s 2021 electricity consumption [70].
   The Greenhouse Gas Protocol (GHG Protocol) offers companies different levels of
commitments to advance their sustainability journey. Emissions from semiconductor
device makers fall into different categories [71]:
   Scope 1 emissions arise directly from fabs, primarily from process gases with
high global warming potential (GWP) that are used during wafer etching, chamber
cleaning, and other tasks; they can also come from high-GWP heat-transfer fluids
that may leak into the atmosphere when fabs use them in chillers. For example, NF3
has a global warming potential that is 17.000 times more potent than CO2 [72]. A
fraction of these potent gases are released to the atmosphere and directly contribute
to global warming.
   Scope 2 emissions arise directly from purchased electricity, steam, heating, and
cooling equipment; the major sources include production tools and facilities/utilities.
   Scope 3 emissions include all other indirect emissions in a company’s value chain;
upstream emissions are those generated by suppliers or their products, while down-
stream emissions are related to the usage of products containing semiconductors.
The number and quantity of materials used for IC manufacturing is steadily growing
with every technology node (e.g., Si wafers, bulk gases, gas precursors, minerals,
chemicals). The upstream production of these materials also leads to greenhouse gas
emissions that must be factored into the emission of the IC chip manufacturing [72].
   Majority of the semiconductor fabrication emissions are from scope 1 process
gases and scope 2 electricity consumption of tools, see Fig. 22.
   Semiconductor companies recognize that many of their customers have set aggres-
sive net-zero targets for their supply chain, see Fig. 23. To address these concerns,
some large semiconductor companies have begun to set ambitious sustainability
commitments.
   Figure 24 shows the comparison of total carbon footprint (or emission) in kgCO2 /
Wafer for the different technology nodes from N28 to N5. The carbon footprint for
294                                                                              A. Moradian
Fig. 22 Majority of the semiconductor fab emissions are from scope 1 process gases and scope 2
electricity consumption of tools [71, 96]
full flow fabrication of an N5 device is almost two times larger than that of an N28
device technology [73].
   Semiconductor companies are also committing to carbon reductions.
. Applied Materials is actively working with customers to look at how to reduce the
  energy used in the fab during the manufacturing of the chips and aims to achieve
  net-zero global emissions for scopes 1, 2 and 3 by 2040 (Applied Materials inc.,
  2023, [87]).
. Lam Research has committed to achieve 100% renewable electricity in 2030, net-
  zero operations (Scope 1 and 2) in 2040 and net-zero Scope 3 emissions by 2050
  (Lam Research inc., 2023, [88]).
Digital Twins for Sustainable Semiconductor Manufacturing                         295
Fig. 23 Many semiconductor companies and their customers have made aggressive net-zero
commitments [71]
Fig. 24 Evolution of total carbon footprint for different technology nodes [73]
296                                                                            A. Moradian
SEMI S23 is the standard [75] that addresses concepts related to energy, utilities, and
materials measurement and use efficiency of semiconductor manufacturing equip-
ment . Semi S23 is intended to be a tool that can be used to analyze energy, utilities,
and materials usage. Other relevant standards include ISO 50001, which provides
guidance for energy management systems—Requirements with guidance for use,
as well as a framework for continues monitoring and improvement (i.e., Plan, Do,
Check, Act) [76]. Another relevant standard is ISO 20140 which provides directions
for automation systems and integration and evaluating energy efficiency as well
as other factors of manufacturing systems that influence the environment [77]. The
German Standards Institute (DIN) requires the Asset Administration Shell (AAS)—a
reference framework for Digital Twins [78]—to be suitable for containing sustain-
ability data and to provide it at the end of a product’s lifecycle for efficient disposal or
recycling. DIN provides a similar concept in the form of a lifecycle record for tech-
nical plants [51]. Further standardizations that specifically address isolated aspects
of sustainability (e.g., energy efficiency) in the field of Industry 4.0 and Digital Twins
is IEC 62,832–3 [79, 80].
The largest use of water (about three-quarters) in a fab is process related, with much of
that being converted to ultra-pure water (UPW) needed for production itself, followed
by the facility scrubber and cooling tower (both about one-tenth), see Fig. 25. Fabs
typically have separate circuits for ultrapure water (UPW), which can be hot and
cold, and lower purity (LP) water. UPW generation is a complex, multistep process
that also consumes significant amounts of power. Most fabs have some level of UPW
reclamation, although rates vary widely among fabs and processes within a fab.
Many processes associated with water usage also contribute to the industry’s carbon
footprint through their consumption of energy. The latest update to S23, SEMI’s
guidance for energy use in the fab, assigns a value of 9 kWh/m3 to generate cold
UPW, and 92 kWh/m3 for hot UPW. S23 does not declare a value for wastewater
management.
Digital Twins for Sustainable Semiconductor Manufacturing                          297
Fig. 25 An estimate of
water usage split in a
semiconductor fab [82]
    Frost and Hua published a study where they tried to quantify spatio-temporal
impacts of the interaction of water scarcity and water use by the global semicon-
ductor manufacturing industry. Their analysis is useful as a benchmark for global
semiconductor industry water withdrawals and may assist OEMs in decisions about
supply chain sourcing. This could also guide semiconductor manufacturers in priori-
tizing locations and time periods to implement water-saving technologies or employ
less water intensive electricity sources. One highlight from this study was that in the
case of semiconductor manufacturing, the most efficient way to reduce overall manu-
facturing water withdrawals is through reduction in fab electricity use. Reductions
in electricity water use can also be achieved by using less water intensive sources of
electricity, such as solar PV and wind, which is especially important during seasons
of higher water scarcity [81].
    The total water drawn from the network for different technology nodes is shown
in Fig. 26.
With the arrival of digital transformation innovations such as generative AI, industry
4.0, autonomous vehicles, and Digital Twins the demand for semiconductor chip
manufacturing continues to grow. As a result, sustainability measures and reducing
the environmental impact will become a key area of focus for manufacturers as well
as governments and local energy regulators. To address this sustainability challenge,
companies need to work individually as organizations and collectively as an industry
to define goals and agree on actions.
   Efforts to address eco-efficiency in semiconductor manufacturing has been
ongoing for the past decade, for instance Higgs et al. 2012 reviewed historic trends in
energy use and environmental impacts of the finished products and identified some
of the key challenges for reducing these impacts going forward [83]. By creating
digital replicas of physical systems, Digital Twins can perform real-time monitoring,
298                                                                                  A. Moradian
Fig. 26 The total water drawn from the network for different technology nodes [73]
analytics, and simulation of energy systems, infrastructure, and ecosystems. This can
aid in resource management, which is key to mitigating climate change [84].
    Using Digital Twins for quantifying carbon footprint, energy, and chemical
consumptions of semiconductor manufacturing equipment was first implemented
in EcoTwin™ [85]. EcoTwin™ is a Digital Twin developed by Applied Materials
that enables on-demand energy and chemical consumption of semiconductor manu-
facturing equipment. By synchronizing data from multiple sensors, and leveraging
domain knowledge and models, detailed reports of carbon footprint and related
consumption parameters from both the fab and sub-fab are automatically generated.
EcoTwin Monitor provides transparency and actionable insights to identify opportu-
nities for sustainable solutions in product and process development [86]. Figure 27
illustrates convergence and aggregation of multiple data streams from subfab and
fab tools to portray a complete picture of consumptions and carbon footprint of the
manufacturing site. Such a digital thread can be extended across multiple fabs and
within as organization and across diverse geographical locations.
    The implementation of EcoTwin is envisioned over a couple of distinct phases. In
its Monitor phase, EcoTwin provides transparency and actionable insights to iden-
tify opportunities for sustainable solutions in product and process development,
see Fig. 28. In the Explore phase, EcoTwin provides a myriad of tools aimed at
consumption predictions, what-if scenarios, and beyond. Ultimately, the manufac-
turers may adopt and implement opportunities for autonomous decision-making and
action-taking steps triggered by this system, at the chamber, tool, or fleet-level. This
Digital Twins for Sustainable Semiconductor Manufacturing                              299
Fig. 28 Overall capability snapshot of EcoTwin Monitor for semiconductor manufacturing carbon
footprint and consumption monitoring platform [86]
Fig. 29 EcoTwin Tool-level dashboard – user can monitor overall impact as well as diving into
each of the subcomponents in fab or subfab
Fig. 30 Chamber-level dashboard with detailed consumptions and carbon footprint contributions
Digital Twins for Sustainable Semiconductor Manufacturing                                     301
Fig. 31 Process recipe consumptions, direct electrical, carbon footprints, and volumetric consump-
tion of chemicals
5 Conclusion
   Moreover, Digital Twins pave the way for workforce development initia-
tives centered around virtual representations of manufacturing equipment, integra-
tion processes, and device design and verification. By leveraging Digital Twins
of virtual fabs, personnel can engage in immersive training experiences and gain
invaluable insights into the intricacies of semiconductor manufacturing.
   In essence, the integration of Digital Twin technology holds the promise of
revolutionizing semiconductor manufacturing, offering a pathway to enhanced effi-
ciency, innovation, and competitiveness in an increasingly dynamic landscape. As
the industry continues to evolve, Digital Twins are poised to play an indispens-
able role in driving progress and unlocking new possibilities on the semiconductor
manufacturing frontier.
References
 1. Grieves M (2005) Product lifecycle management: the new paradigm for enterprises. Int J Prod
    Dev 2(1–2):71–84
 2. Shafto M, Conroy M, Doyle R, Glaessgen E, Kemp C, LeMoigne J, Wang L (2010) Modeling,
    simulation, information technology and processing roadmap, NASA
 3. Grieves M (2015) Digital Twin: manufacturing excellence through virtual factory replication
 4. Negri E, Fumagalli L, Macchi M (2017) A Review of the roles of Digital Twin in CPS-based
    production systems. Proc Manuf 11:939–948
 5. Tao F, Zhang H, Liu A, Nee AYC (2019) Digital Twin in industry: state-of-the-art. IEEE Trans
    Ind Inf 15(4):2405–2415
 6. Tao F, Zhang M, Liu Y, Nee A (2018) Digital Twin driven prognostics and health management
    for complex equipment. CIRP Ann 67(1):169–172
 7. DTC, DIgital Twin Consortium (2020) [Online]. Available: https://www.digitaltwinconsort
    ium.org/hot-topics/the-definition-of-a-digital-twin/
 8. Feng S, Jones A, Shao G (2023) ISO 23247–1 Automation systems and integration—Digital
    twin framework for manufacturing—Overview and general principles, International organi-
    zation for standardization. In: Data requirements for digital twins in additive manufacturing,
    ASME manufacturing science and engineering conference, New Brunswick, NJ, US
 9. Digital Twin Consortium (DTC), Capabilities Periodic Table (CPT) (2022) [Online]. Available:
    https://www.digitaltwinconsortium.org/initiatives/capabilities-periodic-table/
10. Warke V, Kumar S, Bongale A, Kotecha K (2021) Sustainable development of smart
    manufacturing driven by the Digital Twin framework: a statistical analysis. Sustainability
    13:10139
11. Dalibor M, Jansen N, Rumpe B, Schmalzing D, Wachtmeister L, Wimmer M, Wortmann A
    (2022) A Cross-domain systematic mapping study on software engineering for Digital Twins.
    J Syst Softw 193:111361
12. Change2Twin Project, Change2Twin—Digital Twin for every manufacturing SME! (2020)
    [Online]. Available: https://www.change2twin.eu/
13. Land I, Niatas J, Serughetti M (2023) What are Digital Twins? A primer on virtual models
    (2023) [Online]. Available: https://www.synopsys.com/blogs/chip-design/digital-twins-sem
    iconductor-industry.html
14. National Academies of Sciences, Engineering, and Medicine (2024) Foundational Research
    Gaps and Future Directions for Digital Twins. Washington, DC: The National Academies Press.
    https://doi.org/10.17226/26894.
15. Henning S, Hasselbring W (2012) The Titan Control center for industrial devops analytics
    research. Softw Impact 7:100050
Digital Twins for Sustainable Semiconductor Manufacturing                                       303
16. Henning S, Hasselbring W, Burmester H, Möbius A, Wojcieszak M (2021) Goals and measures
    for analyzing power consumption data in manufacturing enterprises. J Data Inf Manag 3(1):65–
    82
17. Seegrün A, Kruschke T, Mügge J, Hardinghaus L, Knauf T, Riedelsheimer T, Lindow K (2023)
    Sustainable product lifecycle management with Digital Twins: a systematic literature review.
    Proc CIRP 119:776–781
18. XMPro inc (2023) [Online]. Available: https://xmpro.com/platform/. Accessed 11 Nov 2023
19. Green G (2023) XMPRO-Digital Twin consortium 2023: a digital twin-based remote operations
    center for wind farms. [Online]. Available: https://www.brighttalk.com/webcast/18347/592103
20. Abdoune F, Ragazzini L, Nouiri M, Negri E, Cardin O (2023) Toward Digital Twin for sustain-
    able manufacturing: a data-driven approach for energy consumption behavior model generation.
    Comput Ind 150:103949
21. Asrai RI, Newman ST, Nassehi A (2018) A mechanistic model of energy consumption in
    milling. Int J Prod Res 56(1–2):642–659
22. Lv J, Tang R, Tang W, Liu Y, Zhang Y, Jia S (2017) An investigation into reducing the spindle
    acceleration energy consumption of machine tools. J Clean Prod 143:794–803
23. Lee W, Kim SH, Park J, Min BK (2017) Simulation-based machining condition optimization
    for machine tool energy consumption reduction. J Clean Prod 150:352–360
24. Seow Y, Rahimifard S (2011) A framework for modelling energy consumption within
    manufacturing systems. CIRP J Manuf Sci Technol 4(1):258–264
25. Karanjkar N, Joglekar A, Mohanty S, Prabhu V, Raghunath D, Sundaresan R (2018) Digital
    Twin for energy optimization in an SMT-PCB assembly line. In: 2018 international conference
    on internet of things and intelligence system (IOTAIS)
26. Walther J, Weigold M (2021) A systematic review on predicting and forecasting the electrical
    energy consumption in the manufacturing industry. Energies 14(4):968
27. He Y, Wu P, Li Y, Wang Y, Tao F, Wang Y (2020) A generic energy prediction model of machine
    tools using deep learning algorithms. Appl Energy 275:115402
28. Sossenheimer J, Walther J, Fleddermann J, Abele E (2019) A sensor reduced machine learning
    approach for condition-based energy monitoring for machine tools. Proc CIRP 81:570–575
29. Lee D, Meyer JS, Touloukian P, Story J, Schneider P, Sartini J (2023) White paper: how Intel®
    automated factory solutions delivers tremendous benefits to semiconductor manufacturing
    environments, Intel
30. TSMC—Machine Learning for Quality Management (2023) [Online]. Available: https://www.
    tsmc.com/english/dedicatedFoundry/services/apm_intelligent_packaging_fab/intelligentFab_
    automation
31. Smart sight: how Micron uses AI to enhance yield and quality (2023) [Online]. Avail-
    able: https://www.micron.com/insight/smart-sight-how-micron-uses-ai-to-enhance-yield-and-
    quality
32. Chen H, Jeremiah S, Lee C, Park J (2023) A Digital Twin-based heuristic multi-cooperation
    scheduling framework for smart manufacturing in IIoT environment. Appl Sci 13(3):1440
33. Rehman A, Ahmed M, Begum S (2023) Cognitive Digital Twin in manufacturing: a heuristic
    optimization approach. In: Artificial intelligence applications and innovations; AIAI 2023. IFIP
    advances in information and communication technology, vol 676
34. Bhalode P, Metta N, Chen Y, Ierapetritou M (2020) Efficient data-based methodology for model
    enhancement and flowsheet analyses for continuous pharmaceutical manufacturing. Comput
    Aided Chem Eng 48(2020):127–132
35. Huang Z, Fey M, Liu C, Beysel E, Xu X, Brecher C (2023) Hybrid learning-based digital twin
    for manufacturing process: modeling framework and implementation. Robot Comput-Integr
    Manuf 82:102545
36. Langlotz P, Klar M, Yi L, Hussong M, Sousa FJ, Aurich JC (2022) Concept of hybrid modeled
    digital twins and its application for an energy management of manufacturing systems. Proc
    CIRP 112:549–554
37. Es-haghi MS, Anitescu C, Rabczuk T (2024) Methods for enabling real-time analysis in digital
    twins: a literature review. Comput Struct 297:107342
304                                                                                  A. Moradian
38. Raissi M, Perdikaris P, Karniadakis G (2019) Physics-informed neural networks: a deep learning
    framework for solving forward and inverse problems involving nonlinear partial differential
    equations. J Comput Phys 378:686–707
39. Scarselli F, Gori M, Tsoi AC, Hagenbuchner M, Monfardini G (2009) The graph neural network
    model. IEEE Trans Neural Netw 20(1):61–80
40. Sanchez-Gonzalez A, Godwin J, Pfaff T, Ying R, Leskovec J, Battaglia PW (2020) Learning
    to simulate complex physics with graph networks. In: Proceedings of the 37th international
    conference on machine
41. Wang J, Huang Y, Chang Q, Li S (2019) Event-driven online machine state decision for energy-
    efficient manufacturing system based on digital twin using max-plus Algebra. Sustainability
    11(18):5036
42. Jia S, Tang R, Lv J, Yuan Q, Peng T (2017) Energy consumption modeling of machining
    transient states based on finite state machine. Int J Adv Manuf Technol 88:2305–2320
43. Dietmair A, Verl A (2009) A generic energy consumption model for decision making and
    energy efficiency optimisation in manufacturing. Int J Sustain Eng 2(2):123–133
44. interTwin Project, "interTwin," (2023) [Online]. Available: https://www.intertwin.eu/about-int
    ertwin/
45. Schmelter S (2023) Trustworthy virtual experiments and digital twins (ViDiT)—Funded
    [Online]. Available: https://www.vidit.ptb.de/home
46. Xiang F, Zhang Z, Zuo Y, Tao F (2019) Digital Twin driven green material optimal-selection
    towards sustainable manufacturing. Proc CIRP 81:1290–1294
47. Davila R, Maria F, Schwark F, Dawel L, Pehlken A (2023) Sustainability Digital Twin: a tool
    for the manufacturing industry. Proc CIRP 116:143–148
48. Ma S, Ding W, Liu Y, Ren S, Yang H (2022) Digital twin and big data-driven sustainable smart
    manufacturing based on information management systems for energy-intensive industries. Appl
    Energy 326:119986
49. Li H, Yang D, Cao H, Ge W, Chen E, Wen X, Li C (2022) Data-driven hybrid petri-net based
    energy consumption behaviour modelling for digital twin of energy-efficient manufacturing
    system. Energy 239:122178
50. Almström P, Andersson C, Öberg A, Hammersberg P, Kurdve M, Landström A, Shahbazi S,
    Wiktorsson M, Windmark C (2017) Sustainable and resource efficient business performance
    measurement systems—The handbook, Mölndal, Mölndal: Billes Tryckeri
51. Miehe R, Waltersmann L, Sauer A, Bauernhansl T (2021) Sustainable production and the role
    of Digital Twins–basic reflections and perspectives. J Adv Manuf Process 3(2):e10078
52. TwinGoals (2020) EIT manufacturing and ‘digital twins’ improving tomorrow’s manufac-
    turing. [Online]. Available: https://eit.europa.eu/news-events/news/eit-manufacturing-and-dig
    ital-twins-improving-tomorrows-manufacturing
53. Chávez CAG, Bärring M, Frantzén M, Annepavar A, Gopalakrishnan D, Johansson B (2022)
    Achieving sustainable manufacturing by embedding sustainability KPIs in Digital Twins. In:
    Winter simulation conference, Singapore
54. Barenji AV, Liu X, Guo H, Li Z (2021) A Digital Twin-driven approach towards smart
    manufacturing: reduced energy consumption for a robotic cell. Int J Comput Integr Manuf
    34(7–8):844–859
55. Kapteyn M, Pretorius J, Willcox K (2021) A probabilistic graphical model foundation for
    enabling predictive digital twins at scale. Nat Comput Sci 1(5):337–347
56. The White House—Statements (2022) Fact sheet: chips and science act will lower costs,
    create jobs, strengthen supply chains, and counter China. [Online]. Available: https://www.
    whitehouse.gov/briefing-room/statements-releases/2022/08/09/fact-sheet-chips-and-science-
    act-will-lower-costs-create-jobs-strengthen-supply-chains-and-counter-china/
57. U.S. Department of Commerce (2024) CHIPS for America Announces $285 million funding
    opportunity for a digital twin and semiconductor chips manufacturing usa institute, U.S.
    Department Of Commerce, [Online]. Available: https://www.commerce.gov/news/press-rel
    eases/2024/05/chips-america-announces-285-million-funding-opportunity-digital-twin
Digital Twins for Sustainable Semiconductor Manufacturing                                      305
Abstract The increasing need for industrial automation is driving the adoption
of robotics, with new developments such as human–robot collaboration through
autonomous mobile robots and collaborative robotic arms. While automation
improves product quality and working conditions and lowers manufacturing costs,
it can also limit manufacturing adaptability. Therefore, when integrating robots in
manufacturing, there is a pressing need to simplify the methods to develop, install,
reconfigure, and operate robot systems to achieve greater adaptability. This is where
the concept of Digital Twin, which replicates the behavior of a complex system in
a virtual environment for analysis and optimization, comes into play. In robotics,
the Digital Twin technology is expected to address the challenges associated with
designing, testing, commissioning, and reconfigurations. This requires a Digital Twin
to accurately represent various dimensions of a robotic system under production
variables. This study characterizes the components of a robot system that need to be
modeled in a Digital Twin to create a trustworthy virtual replica of a physical robot
system. A Digital Twin of this kind can be utilized throughout the lifecycle of the
physical robot installation across various use cases.
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2024          307
M. Grieves and E. Hua (eds.), Digital Twins, Simulation, and the Metaverse, Simulation
Foundations, Methods and Applications, https://doi.org/10.1007/978-3-031-69107-2_13
308                                                                    A. A. Malik et al.
1 Introduction
Modern manufacturing systems are complex, with facilities evolving into large
networks of data-connected mechatronic components [1]. The complexity of these
systems stems from the quantity of information that spans various lifecycle phases,
including design, development, commissioning, operations, and end-of-life [2, 3].
The elevated complexity and interconnectivity make the lifecycle management
of modern-day manufacturing systems more challenging. Conversely, demand is
increasing for manufacturing systems to possess resilience and adaptability [4].
Addressing these challenges in complex scenarios requires enabling smart manu-
facturing through digitalization, data connectivity, and the integration of machine
learning [5]. Smart manufacturing, besides resilience, can bring cost reduction,
enhance workers’ well-being, and result in a better return on investment (ROI).
    Industry 4.0, or the fourth industrial revolution, is the net sociotechnological
impact of infusing emerging technologies such as additive manufacturing, machine
learning, robotics, simulations, and the Internet of Things in products and their manu-
facturing systems. Advanced robotic automation stands out as one of the enablers
for Industry 4.0 [6]. Modern installations strive to make robotic automation more
flexible, adaptable, safe, and cost-effective than traditional robotics implementa-
tions. However, flexible approaches are lacking in developing plug-and-play hard-
ware, programming the robots, control program generation, task scheduling, layout
planning, safety assessment, and alignment with production plans.
    Advancements in virtualization, sensing technologies, and computing power facil-
itate the realization of Digital Twins (DTs), which enable the testing and validation
throughout the design, development, and control phases of a complex system in a
virtual space. Different scientific domains increasingly recognize the potential value
of DTs for managing complexity in areas such as manufacturing, transportation,
aircraft, and space missions [7]. Manufacturing customization and reconfigurability
are vital domains to manage through DTs [7].
    Computer models provide a means to shorten the time needed to design, redesign,
and deploy robot systems. Computer-based virtual models of physical systems can be
beneficial for testing and validating the production before implementation [8]. While
this method is consistent with traditional virtual modeling, the emerging “lifecycle”
approach and real-time communication between physical and virtual systems are
pivotal concepts of DTs [9].
    Many studies have documented the potential advantages and relevance of
employing DTs for robot systems [10–12]. It has also been observed that devel-
oping a trustworthy virtual replica of a robot system is time-consuming and
demands advanced engineering skills and investment in different engineering soft-
ware. Creating and deploying a DT of a robot system should be structured, simplified,
and standardized to realize the needed ROI. It requires identifying the components of
a robot system that are relevant to the purpose of its DT. Moreover, the flexibility of
the DT itself is critical to ensure that the DT can effortlessly be adapted to evolving
circumstances.
Digital Twins for Robot Systems in Manufacturing                                     309
   This chapter presents the importance of DTs in robotic installations within manu-
facturing systems. The components outlined in a DT of a robot system can assist
researchers and practitioners in developing cost-effective, modular, and flexible DTs,
thereby improving the resilience of robot installations. This is an essential step toward
achieving adaptable manufacturing systems.
   The key contributions of this chapter are to:
1.   Present the components of a DT of a robot system for flexibility
2.   Examine the lifecycle phases of a robot system that a DT can support
3.   Apply DTs in robot systems in manufacturing settings
4.   Present use cases that demonstrate the utilization of DTs in robot systems.
The continuous drive to shorten product life cycles emerges as a significant transfor-
mation in the contemporary business landscape [13]. Emerging sociotechnological
trends require shorter product development and launch timelines. In this setting,
manufacturing companies leverage emerging hardware and software technologies
and their potential opportunities [14, 15].
    Aside from the rapid pace of changes, manufacturers face a shortage of skilled
workers. The recent global exposure to the COVID-19 pandemic also displayed
widespread disruptions in supply chains [16] partly because of a shortage of workers
due to social distancing measures. Research studies identified that future factories
could better address such challenges by adopting modular, flexible, and human-
friendly automation solutions [17].
    A way to develop human-friendly automation solutions is through flexible, collab-
orative robots. Technologies that facilitate the swift validation of new manufacturing
strategies are also needed. Therefore, future manufacturing systems must not only
be repurposable but also be designed, developed, commissioned, and reconfigured
at a significantly faster pace [18].
    DTs can be utilized to address the resilience requirements within a manufac-
turing system. For example, DTs can help reduce the time required to validate
new manufacturing strategies, generate automation programs, and provide main-
tenance support. Additionally, DTs can harness real-time and historical data to offer
insights for process optimization. Such assistance can potentially enhance the level
of resilience that a manufacturing system can provide.
310                                                                     A. A. Malik et al.
Automation describes assigning physical and cognitive tasks to machines and soft-
ware to boost production and decrease human effort [19] In manufacturing, adopting
automation brings advantages such as enhancing workplace safety, efficiency, quality,
and cost-effectiveness [20, 21]. However, this often comes at the cost of reduced
production flexibility. At the heart of industrial automation lies industrial robots.
The subsequent sections elaborate on the diverse types of robots employed in manu-
facturing facilities. Figure 1 shows various industrial robot types, including spher-
ical, SCARA (Selective Compliance Assembly Robot Arm), delta, Cartesian, and
humanoid robots [22]. The robot selection for a specific task is based on the nature of
the tasks to be automated, available space, financial constraints, and process-related
considerations. While these robots enhance manufacturing efficiency, their applica-
tions are limited in certain operations, such as assembly, which only constitutes 7.3%
of robotic use [23].
Robots are the predominant force driving the industrial automation of physical tasks
[22, 24]. These robots, characterized by fixed positioning, operation within enclo-
sures, and time-consuming reconfiguration processes [25], fall into the category of
fixed automation solutions. They can help achieve high production volumes but
must strictly be separated from human interaction. They also demonstrate limited
flexibility [26]. Industrial robots have proven successful in various sectors, including
automotive, medicine, food, and electronics manufacturing [27]. The primary reason
for the unsuitability of robots in assembly is human safety and the challenges of their
reconfiguration [28].
Modern industrial robots are lighter, portable, easy to program, and safe. This change
can meet the need for flexibility in terms of mobility, capability, and capacity [29].
These robots, designed for collaboration and coexistence with humans (Fig. 2), are
commonly known as “cobots” or collaborative robots [30]. A collaborative robot
can be defined as a mechanical device intended for direct physical interaction with
humans, a concept first introduced by Colgate [31] and further developed by Kruger
[32]. Cobots allow humans and robots to work together to harness the strengths
of humans and machines. This concept, often called lean automation, exists at the
convergence of human flexibility and machine efficiency [33].
    Literature showcases diverse applications of cobots, spanning pick-and-place
operations, assembly tasks, welding, inspection processes, and packing [34]. More-
over, cobots have been explored as a viable solution for rapidly repurposing factories
in emergencies [17]. The predominant use of cobots has been in manufacturing small
components such as those assembled into electronics, appliances, and electronic actu-
ators [28, 35]. With the advancement of sensing and safety technologies, cobots are
being considered to automate large and heavy components.
in manufacturing settings [40]. Their adaptability and versatility make them well-
suited for tasks requiring interactions with humans. Beyond manufacturing, AMRs
have found applications in warehouses, military operations, healthcare, search and
rescue missions, security, and home environments [38]. This versatility underscores
the potential of mobile robots to automate operations in diverse fields.
   Different robot types have standard features such as mechanical multijointed
reprogrammable actuators, end-of-arm tooling, machine vision, positioning tech-
nology, and control programs. Adaptability is recognized as necessary for most
modern-day robots. The following section presents a typical physical architecture of
a robot installation.
also facilitate the mobility of a robotic arm within the robot system. Furthermore,
collaborative robots must comply with the ISO19649 standard [41], which estab-
lishes terms related to mobile robots operating on solid surfaces and engaging in
industrial robot applications. To effectively employ the concept of DTs, it is vital to
model most, if not all, of these features of robot systems in their digital models.
A robotic system undergoes a comprehensive life cycle, commencing with its design
and concluding at its end-of-life stage. Correspondingly, its DT follows a parallel life
cycle, adapting to various scenarios and system evolution throughout the lifecycle
(Fig. 5). The subsequent section describes the functions of a DT across multiple
stages in a robotic system’s life cycle.
The results derived from the design phase provide information for developing the
components of the physical system. The developed system then moves to the commis-
sioning stage. In the case of a robot system, this stage may entail the creation of work-
stations, fixtures, feeding devices, and other hardware elements. The Bill of Materials
(BOM) and Bill of Processes (BOP) can be generated, guiding the development of
the physical system. Throughout this phase, the connection between the physical
systems and their DTs can be established by linking the DT to an actual controller or
programmable logic controller (PLC) to identify potential errors. This methodology
is analogous to virtual commissioning (VC). VC, or hardware-in-the-loop simula-
tions, reduces development time by facilitating virtual testing and integration well
before actual commissioning. The physical robot can be live connected with its DT,
allowing it to execute tasks as designed in the DT.
purpose. This CAD data can be obtained directly from the equipment manufacturer,
often in standard exchangeable formats such as STEP (Standard for the Exchange
of Product Data) [45]. Robot manufacturers offer CAD models of their robots, and a
similar practice is followed by manufacturers of related equipment, such as grippers,
fixtures, feeders, and tables. Furthermore, many simulation tools feature a built-in
library of proprietary and generic factory resource CAD models.
    A critical step in preparing the CAD data is creating an assembly file and consol-
idating the individual CAD models of various devices and equipment. The virtual
assembly model must represent the complete physical robot system being investi-
gated. Each component can be assigned material properties and visualization to aid
in subsequent analyses. This file can be exported to various exchangeable formats,
with STEP being the most common standard format.
The CAD data can be imported into a continuous simulation environment. Creating a
dynamic simulation starts with defining the kinematics of each active resource within
the system. It involves specifying position and location constraints, joint types, joint
limits, and velocity limits. For example, a gripper may need to be defined for its
318                                                                  A. A. Malik et al.
motion kinematic joint types, limits, and action poses. The visualization/simulation
of a DT is achieved through three steps: (1) creating the simulation model of a
robot system along with its operation sequences, akin to a Gantt chart, (2) an event-
driven continuous simulation that runs for a pre-determined time and controlled by
an internal logic engine, and (3) the simulation is controlled through signals from a
virtual PLC and other emulators. This simulation becomes the primary component of
the DT for visualization, experimentation, and analysis. After the simulation, it can
perform analyses (e.g., collision detection, layout assessment, cycle time estimates)
and optimizations. Numerous proprietary tools are available to create this type of
simulation, while open-source engines can also be utilized.
PLCs serve as industrial computers for programming and monitoring industrial robot
systems. A critical step in commissioning a robot-based manufacturing system is
creating and validating the automation program. Usually, this program is created
later in the development stages. Developing, testing, and validating the automa-
tion program in a virtual space, along with process simulation, enhances the relia-
bility of the system’s performance. Each PLC has its programming tool, and open-
source program development tools are also available. To ensure interoperability,
PLC programs follow the IEC 61131 standard [46]. The IEC 61131-3, developed
by the International Electrotechnical Commission (IEC), sets the standard for PLCs’
syntax, semantics, and interoperability. The developed programs are downloaded
onto a virtual PLC and interfaced with the simulation.
A robot system includes sensors, actuators, feeders, fixtures, and other mecha-
tronic elements (Fig. 3). Behavioral modeling of these devices enables an accurate
virtual model of the entire system. The Functional Mock-up Unit (FMU) is a tool-
independent, free standard crafted for dynamic model exchange and co-simulation.
FMUs define a container and an interface for sharing dynamic simulation models
through a combination of Extensible Markup Language (XML) files, binaries, and
C-code. Both commercial and open-source tools are accessible for simulating the
behavior of each device and interfacing it with the process simulation.
Digital Twins for Robot Systems in Manufacturing                                     319
The evaluation of production variables based on historical statistical data is not always
required in the process modeling of a robot system. However, it is a component that
can be added to a DT of a robot system for production-related analysis. In this phase,
production-related parameters and throughput specifications are defined. Since these
analyses are done in a stochastic simulation, a different tool is often required to
perform stochastic modeling and interface it with the existing continuous process
simulation model. This model simulates cycle time, startup time, setup time, potential
failure scenarios, repair and maintenance requirements, shifts, worker allocation, and
other relevant factors with statistical probabilities. An example of such a simulation
is shown in Fig. 7, where six human–robot packaging stations are shown, and the
simulation presents the operational time and waiting time for each resource. Discrete
simulations can run thousands of trials based on probabilistic distributions derived
from historical data. This analysis enables the prediction of the throughput of a robot
system under the design variables.
allows end users to interact with the robot system in a way that is similar to its actual
application. HMIs are designed following the ISA 101 standard, and validating them
may require a DT model for thorough validation. A virtual HMI may be accessed on
a computer screen or a handheld computing device communicating with the process
simulation. The HMI developed during this step is downloadable to real HMIs for
practical field applications.
In robot system DTs, real-world data are integrated into the simulation. This integra-
tion enables continuous assessments under variable conditions, enhancing the accu-
racy and responsiveness of the DT. System performance data, including logs and
alerts, can be stored in a data repository for ongoing evaluations. Various sensors,
tailored to specific requirements, can be employed in robotic systems to log parame-
ters such as robot joint positions, machine vision data, collision events, task comple-
tion status, safety breaches, and cycle time. As exemplified by [1], a robot assembly
system is connected to a cloud data repository through an internet-based router for
data logging. It is an HRC assembly cell using a UR-5 robot. The performance-
related logs from the assembly cell are stored and used in the simulation for design
optimization (Fig. 8). These performance factors include the idle time for both the
operator and the robot, human operator safety (collision occurrences), and completed
cycle counts. The recorded data is then used for layout and robot path optimization
in the simulation.
    Data management organizes data into meaningful information, focusing on
creating “Golden datasets” that undergo cleaning, transformation, validation, integra-
tion, and standardization. The data management lifecycle encompasses key stages:
stakeholders. The DT models can also simulate the safety training scenarios within
a virtual environment to enhance learning and preparedness.
Studies have found that robotic automation projects frequently exceed the initially
projected timeline [52]. This duration can be further extended, particularly with
more complex tasks such as assembly and battery pack manufacturing. Challenges
that emerge unexpectedly in the planning stage, impacting the project timeline,
include issues related to process balancing, task scheduling, feeding methods,
fixtures required, the necessity for multiple grippers, and safety complications.
This prolonged integration timeframe is attributed to the nature of robot operations.
Robots must coordinate their movements with other hardware (machine tools, equip-
ment, end-of-arm tooling), peripherals (vision systems, force sensors), and humans,
contributing to increased integration and operational complexity.
    Various methods and frameworks have been documented for developing robot
systems [53–55]. However, there is a growing need for novel approaches that focus
on minimizing the time and effort required for integration, validation, and reconfig-
uration. DT prototypes can offer a high-fidelity and trustworthy digital model of the
real system to reduce the chances of any errors that may arise at a later stage.
Robot programming involves defining the paths, actions, and logical procedures for
robots to perform the assigned tasks. Industrial robotic applications often require
significant expertise and effort. Despite the promise of more accessible program-
ming in the latest robots, manually programming complex robot paths remains
time-consuming. With the continuous desire for customization and changing market
dynamics, robot programming is not a one-time activity in their operational life. The
needed flexibility and adaptability require easy ways of programming the robots.
   To tackle this challenge, a DT can facilitate the intuitive development of robot
programs offline within a graphical environment (Fig. 9). The robot program can
be seamlessly downloaded from the DT directly to the connected robot by virtually
testing the desired operation, complete with defined robot trajectories and logic. This
approach streamlines the programming process, leading to a significant enhancement
in efficiency during robot deployment.
Digital Twins for Robot Systems in Manufacturing                                   323
Fig. 9 Generating robot program and controlling robot arm through its DT [47]
collisions, and ensure a safe working environment. The test can define a range of
points from which a robot can reach a selected set of locations.
   Other considerations may include designing the layout with flexibility and scal-
ability to accommodate future changes and expansions. Modular design approaches
and flexible workspace configurations enable easier reconfiguration and adaptation to
evolving production needs. Another important consideration is ensuring the seamless
integration of robotic installations with existing manufacturing systems, including
conveyor systems and automated guided vehicles. Coordinating the layout design
with other production processes (inventory management systems) is essential to
optimize workflow integration and synchronization.
Fig. 11 Sketch shows the assembly process sequence in a robotic assembly process
326                                                                   A. A. Malik et al.
Robot systems with multiple robotic arms, sensors, actuators, feeders, and hard-
ware devices typically operate based on a logical control program. Traditionally,
the development and verification of this logical program are carried out during the
commissioning phase. The validation and error-proofing procedures can be time-
consuming and may uncover errors that could set the project back to earlier stages.
However, leveraging a DT allows for a more streamlined approach. The control
program can be generated directly from the simulation model and verified several
times during the system development. This generated program is then downloaded to
a PLC. The advantage of this approach is that, with each change in the system design,
a new control program can be rapidly created, validated within the DT simulation,
and then efficiently downloaded to the PLC. This dynamic adaptability enhances the
efficiency and agility of the robot system, reducing the time and effort traditionally
associated with the commissioning process.
This section presents two use cases from industrial practices that exemplify the
practical application of DTs in robot systems.
This case exemplifies the creation and utilization of a DT for a collaborative mobile
robot aiming to automate assembly tasks in the manufacturing of wind turbines.
Applying a mobile robot assistant for hybrid automation of wind turbine manufac-
turing is expected to give benefits such as reduced production costs, enhanced product
Digital Twins for Robot Systems in Manufacturing                                     327
quality, and improved working conditions. The DT technology was employed for the
commissioning and reconfiguring the mobile robot, contributing to expedited design
and validation.
   The growing demand for higher rating capacity in wind turbines has driven the
production of larger generators, extended blades, and taller towers. The assembly
of these turbines entails the manipulation of large-sized components. Traditional
automation is impractical due to frequent design alterations and diverse tasks,
rendering it labor-intensive. However, human capabilities fall short in managing
the assembly of these substantial components, resulting in extended lead times and
increased overall costs for sustainable energy. To address these automation chal-
lenges, the case presents a solution consisting of a mobile robot equipped with
a robotic arm that uses laser beams to assist operators in precisely positioning
assembly components. This approach streamlined the assembly process and reduced
the assembly time.
   The simulated robot was connected to the physical robot in real time using an
Ethernet connection (see Fig. 12). The robot’s IP address was utilized in the simula-
tion to establish this connection. A post-processor for Doosan Robotics was used. A
robot post-processor defines how robot programs are generated for a specific robot
controller.
   During the design phase, a comprehensive workspace with the robot system was
modeled in simulation-based DT. The DT model served as a design validation tool for
each reconfiguration. During the development phase, the DT model was a reference
for programming the robot. Assembly locations were extracted from the DT model
and encoded in the robot program. Vision tests and safety risk assessments were
performed using the DT. The DT was operational with the robot system to verify the
changes and robot programming.
used to create and select the elements of the assembly system according to produc-
tion requirements and its integration with the overall system. Dynamic simulation
facilitated a quantitative assessment and a business value proposition for the proposed
solution. The Task Simulation Builder in TPS was used to model human tasks for
ergonomic assessments.
    During the operational phase, using data loggers, the DT transitioned from manual
data syncing to automated and real-time data syncing, enhancing its utility for error
identifications. The DT proved valuable in dynamic task distribution based on task
complexity rating and event-driven simulation. Other benefits included intuitive robot
programming to reduce manual efforts, human safety assessments, the generation of
data logs for critical actions through sensor integration, and the incorporation of
artificial intelligence to enable the system to self-learn and make decisions based on
past experiences.
Interoperability stands out as a critical challenge in the current state of creating DTs
for robot systems. The various aspects of a robot system (geometry, kinematics,
robot program, automation logic, etc.) can be modeled using different tools (e.g.,
Robot Operating System (ROS) [57], OpenPLC [58], Unreal Engine [59], etc.).
However, exchanging data between these tools is often complicated and sometimes
impossible. Additionally, commercial tools from one vendor may not exchange data
with those from another. Consequently, there’s a pressing need for standardized,
exchangeable data formats for DT devices and assets, particularly their connectivity
with open-source tools. Initial standards developments such as the IEC 63278—
Asset Administration Shell [60] have been made. Still, they are at an early stage
Digital Twins for Robot Systems in Manufacturing                                    329
of development, and their widespread adoption needs to address their suitability for
various scenarios.
   Creating and operating a DT necessitates investment in software tools, human
resources, data processing ability, and communication technologies. However, it is
essential to recognize that not every robot installation requires a DT. Fundamentally,
a DT benefits robotic systems as it facilitates flexibility, reconfigurability, seamless
deployment, predictive maintenance, and safe human–robot interaction, among other
benefits. Assessment methods must be available to identify the need and value of a
DT in a given context and asses the ROI of creating and using a DT.
   In manufacturing systems, hardware modularization involves incorporating
parallel modules for capacity or capability adjustments. Combining flexible automa-
tion with human skills, this modular approach facilitates reconfigurability and agility.
Achieving a high degree of customization involves adding, replacing, or eliminating
modules. A similar modular approach is needed in DTs. Given that the current
approach to developing DTs is time-consuming, the creation of libraries of modular
DTs of assets and information blocks can pave the way for the formation of reusable
and exchangeable DTs.
   Flexible robot installations face the challenge of effective and dynamic task
scheduling. Task scheduling for robots involves the creation of an optimal schedule
within a system. This schedule specifies which robot is assigned to each task and
how the tasks will be processed. Dynamic task allocation goes a step further, encom-
passing the planning of automation processes and allowing robots to adapt during
operations. To address this challenge, a comprehensive approach that integrates
event-based logical simulation, probabilistic analysis, and statistical data analysis
into a DT model is needed. This integration can enhance the DT’s capability to
facilitate dynamic and efficient task scheduling, ensuring adaptability to changing
operational needs.
11 Conclusion
requires the creation of a reliable and comprehensive digital representation that accu-
rately models the elements and dynamics of the observed robot system. Flexibility
remains critical in ensuring the DT can seamlessly adapt to changing circumstances.
This chapter underscores the significance of DTs in manufacturing robotic cells and
outlines the essential criteria for developing these DTs.
    The interoperability challenge arises in creating DTs for robot systems, as
exchanging data between different tools proves difficult. Standardized and exchange-
able data formats are essential, emphasizing compatibility with open-source tools.
While investing in DT tools and resources is crucial, it’s important to note that
not every robot system requires a DT. Modularization is critical in manufacturing
and DT development, offering flexibility and reconfigurability. Flexible robot instal-
lations encounter challenges in dynamic task scheduling, which can be addressed
by integrating logical simulation and statistical analysis into DT models, ensuring
adaptability to changing operational needs.
Disclaimer This research was conducted through the support of a NIST cooperative agreement
[60NANB23D234]. Specific commercial products and systems are identified in this paper to facil-
itate understanding. Such identification does not imply that these software systems are necessarily
the best available for the purpose. No approval or endorsement of any commercial product by NIST
is intended or implied.
References
 1. Malik AA, Brem A (2021) Digital twins for collaborative robots: a case study in human-
    robot interaction. Robot Comput Integr Manuf 68:102092. https://doi.org/10.1016/j.rcim.2020.
    102092
 2. Rios J, Bernard A, Bouras A, Foufou S (2017) Product lifecycle management and the industry
    of the future. In: 14th IFIP WG 5.1 international conference. Springer, Seville, Spain
 3. Morgan J, Halton M, Qiao Y, Breslin JG (2021) Industry 4.0 smart reconfigurable manufacturing
    machines. J Manuf Syst 59:481–506. https://doi.org/10.1016/j.jmsy.2021.03.001
 4. Yazdani MA, Khezri A, Benyoucef L (2022) Process and production planning for sustainable
    reconfigurable manufacturing systems (SRMSs): multi-objective exact and heuristic-based
    approaches. Int J Adv Manuf Technol 119:4519–4540. https://doi.org/10.1007/s00170-021-
    08409-0
 5. Shao G, Helu M (2020) Framework for a digital twin in manufacturing: scope and requirements.
    Manuf Lett 24:105–107. https://doi.org/10.1016/j.mfglet.2020.04.004
 6. Advanced Manufacturing (2022) National strategy for advanced manufacturing
 7. Malik AA (2023) Simulation based high fidelity digital twins of manufacturing systems: an
    application model and industrial use case. In: Winter simulation conference 2023
 8. Masood T, Weston R (2011) An integrated modelling approach in support of next generation
    reconfigurable manufacturing systems. Int J Comput Aided Eng Technol 3:372–398
 9. Bilberg A, Malik AA (2019) Digital twin driven human–robot collaborative assembly. CIRP
    Ann 68:499–502. https://doi.org/10.1016/j.cirp.2019.04.011
10. Maruyama T, Ueshiba T, Tada M, Toda H, Endo Y, Domae Y (2021) Digital twin-driven human
    robot collaboration using a digital human 1–46. https://doi.org/10.3390/s21248266
11. Liu S, Wang XV, Wang L (2022) Digital twin-enabled advance execution for human-robot
    collaborative assembly. CIRP Ann 71:25–28. https://doi.org/10.1016/j.cirp.2022.03.024
Digital Twins for Robot Systems in Manufacturing                                                 331
12. Yao B, Xu W, Shen T, Ye X, Tian S (2023) Digital twin-based multi-level task rescheduling
    for robotic assembly line. Sci Rep 13:1–20. https://doi.org/10.1038/s41598-023-28630-z
13. Masood T, Kern M, John Clarkson P (2021) Characteristics of changeable systems across value
    chains. Int J Prod Res 59:1626–1648. https://doi.org/10.1080/00207543.2020.1791997
14. Giones F, Brem A (2017) From toys to tools: the co-evolution of technological and
    entrepreneurial developments in the drone industry. Bus Horiz 60:875–884
15. Stadnicka D, Antonelli D (2019) Human-robot collaborative work cell implementation through
    lean thinking. Int J Comput Integr Manuf 32:580–595. https://doi.org/10.1080/0951192X.2019.
    1599437
16. Malik AA, Masood T, Kousar R (2020) Repurposing factories with robotics in the face of
    COVID-19. Sci Robot 5, (2020) https://doi.org/10.1126/scirobotics.abc2782
17. Malik AA, Masood T, Kousar R (2021) Reconfiguring and ramping-up ventilator production
    in the face of COVID-19: can robots help? J Manuf Syst 60:864–875. https://doi.org/10.1016/
    j.jmsy.2020.09.008
18. Anumbe N, Saidy C, Harik R (2022) A primer on the factories of the future
19. Nof SY (2009) Springer handbook of automation. Springer, Berlin
20. Boy GA (2018) From automation to interaction design. Handb Hum-Mach Interact 431–443.
    https://doi.org/10.1201/9781315557380-22
21. Malik AA, Masood T, Management E, Brem A (2024) Intelligent humanoid robots in manu-
    facturing: addressing the worker shortage and skill gaps in assembly cells. https://doi.org/10.
    1145/3610978.3640765
22. Kurfess TR (2004) Robotics and automation handbook
23. Shneier MO, Messina ER, Schlenoff CI, Proctor FM, Kramer TR, Falco JA (2015) Measuring
    and representing the performance of manufacturing assembly robots
24. Andersson SKL, Granlund A, Hedelind M, Bruch J (2020) Exploring the capabilities of indus-
    trial collaborative robot applications. Adv Transdisciplinary Eng 13:109–118. https://doi.org/
    10.3233/ATDE200148
25. Pedersen MR, Nalpantidis L, Andersen RS, Schou C, Bogh S, Krüger V, Madsen O (2016)
    Robot skills for manufacturing: From concept to industrial deployment. Robot Comput Integr
    Manuf 37:282–291
26. Ji S, Lee S, Yoo S, Suh I, Kwon I, Park FC, Lee S, Kim H (2021) Learning-based automation of
    robotic assembly for smart manufacturing. Proc IEEE 109:423–440. https://doi.org/10.1109/
    JPROC.2021.3063154
27. Buerkle A, Eaton W, Al-Yacoub A, Zimmer M, Kinnell P, Henshaw M, Coombes M, Chen
    WH, Lohse N (2023) Towards industrial robots as a service (IRaaS): flexibility, usability, safety
    and business models. Robot Comput Integr Manuf 81:102484. https://doi.org/10.1016/j.rcim.
    2022.102484
28. Malik AA, Bilberg A (2019) Complexity-based task allocation in human-robot collaborative
    assembly. Ind Robot 46:471–480. https://doi.org/10.1108/IR-11-2018-0231
29. Malik AA, Bilberg A (2019) Developing a reference model for human–robot interaction. Int J
    Interact Des Manuf 13:1541–1547. https://doi.org/10.1007/s12008-019-00591-6
30. Rahman SMM, Sadrfaridpour B, Wang Y, Walker ID, Mears L, Pak R, Remy S (2020) Trust-
    triggered robot-human handovers using flexible manufacturing, 1–8
31. Edward J, Wannasuphoprasit CW, Peshkin MA (1996) Cobots: robots for collaboration with
    human operators. In: International mechanical engineering congress and exposition, Atlanta
32. Krüger J, Lien TK, Verl A (2009) Cooperation of human and machines in assembly lines. CIRP
    Ann Manuf Technol 58:628–646
33. Malik AA, Bilberg A (2019) Human centered lean automation in assembly. Procedia CIRP
    81:659–664. https://doi.org/10.1016/j.procir.2019.03.172
34. Ajoudani A, Zanchettin AM, Ivaldi S, Albu-Schäffer A, Kosuge K, Khatib O (2018) Progress
    and prospects of the human–robot collaboration. Auton Robots 42:957–975. https://doi.org/
    10.1007/s10514-017-9677-2
35. Povlsen K (2023) A new generation of robots can help small manufacturers. https://hbr.org/
    2023/11/a-new-generation-of-robots-can-help-small-manufacturers
332                                                                               A. A. Malik et al.
1 Introduction
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2024               333
M. Grieves and E. Hua (eds.), Digital Twins, Simulation, and the Metaverse, Simulation
Foundations, Methods and Applications, https://doi.org/10.1007/978-3-031-69107-2_14
334                                                          J. F. Macfarlane and I. Babur
radio communication devices, are collecting vast amounts of data on vehicle loca-
tion, speed, surrounding environments, and the health of components. This data is
then analyzed to derive insights into vehicle dynamics, such as driving patterns,
congestion levels, and critical braking events.
    At the same time, the infrastructure itself is becoming smarter. Sensors and
cameras installed across various points enable remote monitoring of traffic condi-
tions, asset status, accidents, and weather changes. Traffic management centers are
using this wealth of information to make operational decisions that enhance response
times and overall efficiency. This shift is replacing older hardware systems with intel-
ligent transportation systems (ITS), which include automated features like adaptive
traffic signals, dynamic tolling, and smart parking meters, all designed to streamline
transportation management.
    Additionally, the surge in data is being captured and analyzed through big data
platforms, hosted on the cloud. These platforms aim to provide a holistic view of
transportation networks by integrating data from both vehicles and infrastructure
sensors.
    A transformation toward Mobility-as-a-Service (MaaS) is taking place. This
model leverages the new digital and communication infrastructure to offer on-
demand services such as ride-sharing, bike-sharing, and microtransit through
consumer-friendly mobile apps. These services are becoming viable alternatives
to private car ownership, particularly in urban areas. Alongside this shift, there is
a growing governmental push toward electrification, including the development of
charging networks and electric fleets, aimed at mitigating the environmental impacts
of internal combustion engines. Digital systems are also playing a crucial role in
facilitating the integration and monitoring of these new electric vehicle demands.
    As a whole, this evolving ecosystem of digital technologies is not only enabling
new transportation services and optimizing infrastructure usage, it is also enhancing
the capacity for comprehensive monitoring and data-driven planning and operations.
This transformation is set to accelerate and will reshape the future of transportation
in the years to come.
various scenarios, understand the cascading effects of network changes, and impor-
tantly, identify unintended consequences of these modifications. Enhanced compu-
tational power and advanced simulation technologies are amplifying the capabilities
of Digital Twins, offering unprecedented insight into the dynamics of transportation.
    The development of effective Digital Twins for regional planning hinges on several
key research areas: defining clear fundamental domain taxonomies, advancing simu-
lation methodologies, integrating data from multimodal transportation sources, and
using artificial intelligence to codify urban complexities. Thoughtfully designed,
Digital Twins have the potential to revolutionize the transportation planning process.
They enable more precise analyses of investment options and policy decisions, offer
the capability to evaluate alternatives swiftly, and provide rapid feedback, all of
which contribute to more agile and effective decision-making processes that enhance
mobility outcomes. Realizing the full potential of Digital Twins will necessitate
strong partnerships among researchers, technology providers, planning agencies,
and private industry.
management. These apps can dynamically reroute travelers to faster routes as they
become available, which, while while potentially helpful in reducing congestion, can
also lead to unintended traffic disruptions in residential areas or lead to ill-informed
behaviors in unusual circumstances, such as during wildfire evacuations.
   The primary objective of regional transportation modeling is to codify these
complexities of disjointed traffic management systems. The aim is to develop
analytics that assist city planners in enhancing transportation options and dynamics
across cities and to design strategies that encourage mode shifts which will be crucial
for the future of traffic management in metropolitan areas.
Simulation tools for transportation systems come in various forms, typically catego-
rized into microsimulation, mesosimulation, and macrosimulation. Our novel plat-
form, Mobiliti, stands out by overcoming traditional computational limitations asso-
ciated with the simulation of large regional road networks. It uses an agent-based,
parallel discrete event simulation framework where each road link acts as an agent.
These agents control the movement of vehicles across their geospatial locations,
managing traffic flow according to constraints such as vehicle capacity, speed limits,
and queue lengths. For instance, in the San Francisco Bay Area, Mobiliti can model
24 h of activity for approximately 7 million personal vehicles and 4 million trucks,
translating to approximately 19 million trips across a network of 1 million road links,
in less than eight minutes—a scale of simulation and computational speed unmatched
by any current system.
    Similar simulation tools limited by computational constraints are forced to reduce
the scale of the simulation. This typically involves decreasing the number of trips
or population size, reducing the size of the road network or simplifying its geom-
etry, or any combination of these approaches. These reductions and simplifications
fail to accurately represent the true dynamics of the metropolitan regions being
studied. In practice, this limitation is particularly critical when considering dynamic
rerouting, which often involves lower-class neighborhood roads that are often omitted
in reduced network models.
    In contrast, planning agencies sometimes use optimization algorithms to under-
stand road network loading. These tools analyze trip demand profiles and assign
trips to the network, aiming to optimize travel times. However, these tools also face
computational challenges, and the traffic assignment models often operate with very
low spatial and temporal fidelity—for example, running optimizations over a static
four-hour period. Despite this, these tools can still require up to 24 h to process and
frequently yield results that significantly underestimate congestion patterns.
Mobiliti: A Digital Twin for Regional Transportation Network Design …                337
The importance of regional transportation models stems from their ability to accu-
rately represent the travel demand profiles of individual cities, a necessity for several
reasons. First, economic and land use impacts are inherently regional. Activities such
as employment, shopping, and recreation generate travel patterns that often begin in
one city and conclude in another within the same region. Consequently, examining a
single city in isolation does not capture critical origin–destination flows and requires
assumptions about boundary conditions for trips that cross these boundaries.
    Moreover, the transportation needs of a city are deeply influenced by regional
developments. Public transportation options and road networks extend across city
limits, creating an interconnected system where traffic conditions in one area
can affect another downstream. Additionally, cities often share significant infras-
tructure such as ports, airports, and major state-run corridors, emphasizing the
interdependence of urban areas.
    Transportation policy and investment decisions, typically made at a regional level,
require insights into travel behaviors across multiple cities. By focusing solely on
individual cities, significant interactions and the use of shared assets may be over-
looked. Regional models address these challenges by providing a comprehensive
view of mobility patterns, capturing the intricacies of how people move within and
between urban areas.
    With projections indicating that 68% of the global population will reside in major
cities by 2050, the relevance of regional models becomes even more pronounced. As
urbanization continues to drive growth in major cities and expands them into future
mega-regions, understanding the complex dynamics of regional transportation will
be crucial for managing congestion and planning sustainable urban environments.
When discussing Digital Twins, it’s important to recognize that they are composed
of three distinct components: a physical system operating in real-time, dynamic data
about this system streamed at various frequencies, and a computationally manage-
able simulation of the system. This chapter explores a Digital Twin specifically
designed to address the complex challenges of managing transportation systems
across metropolitan regions.
    The Digital Twin Consortium defines a Digital Twin as “a virtual representa-
tion of real-world entities and processes, synchronized at a specified frequency and
fidelity”. This emphasizes the Digital Twin’s role in mirroring real-world systems
and processes accurately and in real-time.
    Contrasting with this is the model-based controller, a concept well-established
in traditional control theory. Model-based controllers use knowledge of physical
laws governing systems to estimate responses to external inputs. This knowledge is
338                                                          J. F. Macfarlane and I. Babur
embedded within a feedback control system, allowing the controller to predict and
optimize system behavior under various conditions. The challenge lies in developing
accurate models that can sufficiently account for all relevant system behaviors.
   In comparison, a Digital Twin goes beyond control. It serves as a comprehensive
virtual counterpart of the system, not only monitoring and optimizing performance
but also enabling scenario testing without impacting the physical system. Digital
Twins thus offer substantial benefits, such as providing valuable insights, facilitating
remote monitoring and control, and enhancing decision-making processes.
   Key differences between a model-based controller and a Digital Twin include:
. A model-based controller uses a simplified model primarily for control, aiming to
  meet specific control requirements. It does not typically mirror the full complexity
  of the physical system. Conversely, a Digital Twin seeks to emulate the physical
  system as closely as possible, incorporating much greater detail and complexity.
. The model in a model-based controller is generally static, with updates based on
  new data being slow or non-existent. On the other hand, a Digital Twin contin-
  uously updates its virtual representation with real-time data from the physical
  system.
. The primary objective of a model-based controller is to calculate optimal control
  inputs to achieve desired performance. In contrast, the Digital Twin aims to
  understand, analyze, and potentially optimize the physical system, providing a
  predictive analytics capability for exploring future states.
   While model-based controllers use simplified models designed specifically for
control algorithms, Digital Twins aim to fully virtualize the physical system and
maintain synchronization with real-time data. This allows Digital Twins to facilitate
a broader exploration of scenarios in the context of the system’s current state. It is
likely that as Digital Twins evolve, model-based control techniques will be integrated
within them, validated against these higher-fidelity representations to further enhance
their utility and effectiveness.
and simulate the complex dynamics of these extensive networks, providing a valuable
tool for understanding and optimizing transportation infrastructure.
bounds. However, due to the necessity of adhering to the most conservative case,
this method often results in missed opportunities for parallel execution and requires
aggressive synchronization.
A crucial aspect of any Time Warp-based parallel discrete event simulation (PDES)
is managing the computation of Global Virtual Time (GVT). GVT is the consensus
timestamp up to which all events are confirmed as correctly processed, serving as a
baseline for system integrity. Maintaining a queue that can revert the system to any
prior state is essential for enabling rollbacks when events are processed out of order,
though this can significantly increase memory requirements.
    To enhance scalability and manage this complexity, a highly asynchronous version
of Time Warp is implemented that eschews blocking collective operations. A primary
challenge in optimizing Time Warp performance is to curb excessive optimism, which
can lead to frequent and resource-intensive rollbacks. Rollbacks are necessitated
when an event, once processed, must be reversed due to subsequent receipt of an
earlier time-stamped event. This also involves rolling back any events that the initial
event might have triggered on neighboring processors.
    The system leveraged by Mobiliti, called Devastator, was developed to run on a
supercomputer at Lawrence Berkeley National Laboratory by Cy Chan, Max Bremer,
et al. It employs a strategy known as the moving time window. This method limits
the execution interval to GVT plus a tunable parameter (w), thus containing the
temporal range within which events can be executed. As GVT progresses, this
window advances, progressively allowing more pending events to be processed.
Devastator automatically adjusts the width of this window based on operational
efficiency metrics.
342                                                          J. F. Macfarlane and I. Babur
Fig. 1 Overview of Mobiliti highlighting the interconnected components used to analyze and
optimize urban mobility
Fig. 2 Mobiliti link speeds for the Los Angeles basin using the 2020 SCAG travel demand model,
visualization by kepler.gl
The ultimate aim is to establish an explorative platform built upon unified data that
represents a single source of truth while enabling Digital Twin scenario development
capabilities. APIs that facilitate seamless access to this curated data and secure inte-
gration of third-party datasets will encourage users to contribute back to the collective
pool. This fosters a beneficial cycle of data sharing and enhancement across domains,
agencies, and jurisdictions.
    Integrated data science tools embedded within the platform can extract reliable,
reproducible insights from the combined data, positioning the Digital Twin as the
authoritative source for interdisciplinary analysis and preventing the loss of insti-
tutional knowledge. By maximizing the value of multi-source data fusion, a shared
data infrastructure enhances both data quality and reliability through scalable data
science techniques.
    For example, predictive models trained on extensive historical data can detect
anomalies in incoming data streams that deviate from expected patterns. These
conflicts among data from different sensor types can be automatically identified and
flagged for review. This multi-source data fusion can potentially mitigate inherent
biases in individual datasets, revealing biases related to underrepresented populations
and usage patterns, thus supporting more equitable decision-making.
    These capabilities, integrated into an automated data observatory that contin-
uously monitors the Digital Twin’s data ecosystem and identifies accuracy issues
through AI/ML validation pipelines, are crucial to the Digital Twin’s performance.
Issues are prioritized based on their impact, with notifications sent to relevant experts
for necessary adjustments. Engaging the relevant experts in the organization and
involving them in model adjustments based on trending data will be essential to the
long-term success of the Digital Twin.
    Given baseline data quality is verified, completeness can be addressed. For the
transportation domain, data coverage varies widely across different communities and
348                                                           J. F. Macfarlane and I. Babur
may be sparse in certain areas or during specific time periods. Figure 4, for instance,
illustrates traffic count data collected by contractors over several years in San Jose.
    In situations where data gaps exist due to limitations in spatial or temporal
coverage, or sensor malfunctions, the comprehensive data repository of the Digital
Twin can be employed to create synthetic data to fill these gaps. Deep learning
techniques like graph neural networks can assist in this case [14, 15].
    However, it is critical to quantify the uncertainty associated with these synthesized
data products. Real-time data can instill confidence by appearing to accurately reflect
reality, yet it is very often subject to anomalies due to hardware issues, collection
methods, and varying penetration rates. By establishing confidence intervals and
error propagation models, downstream users can assess the reliability of the data
and make informed decisions regarding its suitability for their workflows. This is
illustrated in Figs. 5 and 6, which showcases the data validation processes using
Mobiliti results.
    These advanced algorithms provide a detailed and comprehensive view of regional
behaviors, integrating various influencing factors such as time of day, seasonal varia-
tions, nearby events, weather conditions, population densities, the built environment,
and the economic landscape. These elements are combined within the Digital Twin
to predict traffic flows, speeds, and congestion patterns at a granular level across the
transportation network.
   The simulation model allows for adjustments in individual parameters to gauge
the sensitivity of output metrics to specific changes in the input conditions. The
accuracy of the simulation engine is validated by comparing its predicted outputs
with real-world ground truth data within the Digital Twin demonstrating a favorable
350                                                                   J. F. Macfarlane and I. Babur
Fig. 7 Vehicle count validation allows the Digital Twin to estimate vehicle counts (left) and speeds
(right) across the full network—colored from red to yellow
alignment between simulated and observed data. Once the simulation fidelity is
established through this rigorous benchmarking, the calibrated engine can be used
to fill missing data gaps by generating synthetically simulated values, as shown in
Fig. 7. This approach, corroborated against available estimates, ensures a continuous,
holistic view of the transportation system’s state.
   For short-term predictions, the simulation engine can model the impacts of vari-
ables such as event attendance, weather forecasts, or disruptions in transit services
on expected traffic conditions. For longer-term planning and policy analysis, it eval-
uates scenarios related to demographic shifts, changes in land use, or housing devel-
opments to estimate their effects on key performance indicators including mobility,
accessibility, equity, safety, and environmental impacts.
   The interplay between data-driven simulation and the central unified data repos-
itory enhances the Digital Twin framework’s capabilities. The simulation engine
continuously improves its accuracy by assimilating the latest real-world observa-
tions and can also augment the data repository in areas showing deficiencies. This
closed-loop data enrichment ensures that the Digital Twin maintains a comprehen-
sive, self-regulating representation of the physical system’s state , enabling informed
and reliable decision-making.
   An integrated multi-source data environment facilitates the discovery of novel
relationships and insights through advanced analytics and machine learning. By
applying deep learning techniques to the consolidated data corpus, previously hidden
patterns and complex correlations are uncovered, enabling more accurate modeling
and prediction of system behaviors, as highlighted in Fig. 7, where the real-time data
and learned behaviors tune the model and overcome data sparsity and penetration
concerns.
   Ultimately, by dismantling long-standing data silos and establishing an integrated
Digital Twin environment, cities can unlock the immense potential of their data
assets, deriving reliable, actionable insights to drive intelligent policies, resilient
infrastructure investments, and equitable resource allocation across all communities.
Mobiliti: A Digital Twin for Regional Transportation Network Design …                           351
Proper alignment of real-world data with simulation models is critical for building
trust in the Digital Twin [11]. Travel demand models, which are intricate and costly
to develop, are typically updated at five year intervals. One of the core advantages
of Digital Twins is their ability to incorporate real-time data to refresh these models
more frequently, reflecting the latest trends in demographics, economic conditions,
and land use.
    Moreover, as the Digital Twin generates a comprehensive virtual representation of
traffic, new insights can emerge from the observed traffic patterns. Unlike traditional
methods that may overlook fine-grained details, Mobiliti enables a detailed analysis
across the entire network. For instance, critical link density, detailed in Fig. 8, is
indicative of where vehicle accumulation peaks during the day and provides valuable
insights that might not be feasible to capture through conventional methods.
    Derived traffic insights highlight links that are consistently over-capacity, shed-
ding light on potential congestion causes. Figure 9 identifies locations of saturation
(left) and temporal congestion (right). This information, coupled with expert anal-
ysis, can lead to innovative infrastructure solutions or policy adjustments that alleviate
traffic issues.
    The ability to simulate and predict traffic under various conditions is another
significant benefit of the Digital Twin. By adjusting simulation parameters, the
model can assess the impact of changes in traffic flow, speed, and congestion across
different times of day, seasons, or in response to specific events. These simulations
are validated against real-world data, ensuring the model’s accuracy and reliability.
    Figure 10 compares Average Daily Traffic (ADT) from Mobiliti with actual
observed data, demonstrating the model’s alignment with real-world conditions. It
also shows a comparison to a popular tool that relies solely on mobile device data.
This benchmarking is crucial for confirming the simulation’s accuracy before using
it to fill data gaps or guide decision-making processes.
Fig. 8 Road link level critical density indicates the point at which accumulation of vehicles occurs
352                                                                 J. F. Macfarlane and I. Babur
Fig. 9 Saturation index for the Bay Area model (left); temporal congestion index for the Bay Area
model (right)
   This dynamic interplay between data-driven simulation and a unified data reposi-
tory empowers the Digital Twin to continuously learn and adapt, enhancing its ability
to provide a detailed and accurate representation of the transportation system. This
ongoing data enrichment ensures that the Digital Twin remains a vital, self-updating
tool for decision support.
   Ultimately, by integrating these capabilities into a cohesive Digital Twin frame-
work, municipalities can harness their data resources more effectively, driving
smarter policies, and ensuring equitable resource distribution across all communities.
Mobiliti: A Digital Twin for Regional Transportation Network Design …              353
While the Digital Twin capability provides powerful new computational assets
for practitioners, it does not reduce the complexity of the network fundamen-
tals. As a consequence, defining performance indices that relate to the core needs
of a city’s population remains an important endeavor. To this end, a new frame-
work for evaluating transportation policies and investments that accounts for social
equity impacts alongside traditional mobility performance metrics has been devel-
oped called Socially Aware Evaluation Framework for Transportation—SAEF [9].
Conventional evaluations often focus narrowly on mobility and fail to sufficiently
consider social outcomes. As such, a socially aware framework can provide a post-
simulation data analysis that incorporates quantitative measures of accessibility,
exposure, and externalities experienced by different social groups, including low-
income, minority, and disabled populations. These social impact measures are inte-
grated with mobility metrics within a multi-criteria decision analysis to enable a
more holistic assessment of transportation projects and policies.
The complexity of city management can lead to various urban issues that signif-
icantly affect the quality of life for residents. The Mobiliti platform is currently
geared toward understanding vehicle traffic dynamics, as traffic congestion is a
prominent concern for future city planning, with direct impacts on public health,
safety, and regional economic health. Accordingly, the platform investigates the
city-level impacts of various vehicle routing strategies to enhance traffic manage-
ment and mitigate negative consequences. Beyond traffic management, it can be
used for a variety of analyses, including alternative mode planning.
   The term ‘socially-aware’ encompasses four measurement themes: neighborhood,
safety, mobility, and environment. To address socially relevant aspects of vehicle
dynamics, themes and indicators are developed beyond simple congestion measure-
ments. These themes, supported by city performance indicators from extensive liter-
ature reviews [8], identify critical factors such as accidents for safety and emissions
for environmental quality. This approach has led to the creation of 24 indicators from
an initial list of over 100, categorized into themes that reflect neighborhood quality,
safety, mobility, and environmental quality.
   The framework offers insights into the trade-offs between social and mobility
objectives, aiding in decision-making processes. This methodology is designed
to support more equitable transportation planning outcomes. Fifteen performance
indices, as depicted in Fig. 11, range from positive (green) to negative (red) outcomes
for four different traffic scenarios in the City of San Jose.
354                                                                 J. F. Macfarlane and I. Babur
Fig. 11 SAEF evaluation indices for the city of San Jose under four traffic scenarios (green indi-
cates a positive outcome while red indicates a poor outcome in comparison across the four traffic
scenarios). VHD vehicle hours delayed, VMT vehicle miles traveled
Hours Delayed (VHD) in Equity Priority Communities (EPC) can be extracted from
the simulation results. Designations of disadvantaged communities vary across cities
and area government groups and may consider a variety of factors such as income
level, population age, and ethnicity. For these analyses, geospatial representations
of these EPC communities were used to extract link level speed and flow data and
provide VMT and VHD metrics. As observed in Fig. 12, three of the four scenarios
have poor outcomes for VHD and VMT in EPCs. Further analytics, shown in Fig. 12,
then demonstrate that a significant number of longer trips (arbitrarily selected as trips
greater than 40 min) that start and end in the city centers of San Francisco and San
Jose are using the EPC road links during their journey. These are considered “pass
through” trips, in which the vehicle has no connection to the specific neighborhood
and are short-cutting congestion by using neighborhood streets on their way to their
destination. Further analysis indicates that these trips are mostly work-related trips
or commute trips and that 60% of these trips were rerouted due to congestion, thus
incurring undue safety and health impacts on lower income communities.
   While the simulation may not exactly mirror reality, it provides critical insights
into traffic phenomena, particularly those influenced by driver behaviors, which are
often challenging for traffic planners to measure [13]. This demonstrates a key value
of using Digital Twins in transportation planning.
Fig. 12 Origins and destinations of San Jose and San Francisco commute trips that pass through
lower income communities thus impacting community health and safety
356                                                                    J. F. Macfarlane and I. Babur
8 Iterative Exploration
A high-fidelity virtual representation that mimics the real system enables the safe
and rapid evaluation of scenarios that are challenging or impossible to test in reality.
Gained insights can inform investment plans and policy decisions. The ultimate aim
is to equip practitioners with tools to configure the Digital Twin to reflect poten-
tial changes in infrastructure or technology and assess their impacts on system
performance.
    Examples include:
. Reducing the number of lanes on an arterial road to create space for protected bike
  lanes and observing the effects on congestion and traffic dynamics influenced by
  dynamic routing.
. Adjusting the Origin–Destination Time (ODT) dataset in the Digital Twin to reflect
  changing conditions such as population growth and demand levels and observing
  the resulting strain on network resources. For instance, in San Francisco, a simula-
  tion where current transit riders were instead assigned a personal vehicle showed
  the significant congestion that would result without existing transit systems
  (Fig. 13-left).
. Introducing disruptive events like natural disasters or major accidents into the
  Digital Twin to assess resilience and response capabilities. For example, a simu-
  lated earthquake in San Jose directed drivers within a certain radius to return
Fig. 13 Example scenarios for iterative exploration: include transit riders as vehicle owners/drivers
(left) and a simple earthquake model in which vehicle drivers return home (vehicle density is
indicated from red to yellow)
Mobiliti: A Digital Twin for Regional Transportation Network Design …                  357
  home, dramatically affecting local traffic and also impacting main arterials far
  beyond neighboring cities (Fig. 12-right).
. Using machine learning within the Digital Twin to develop predictive models.
. Engaging stakeholders to define scenarios of interest and exploring consequences
  to foster an intuitive understanding of existing conditions.
   While simulations within the Digital Twin may not always perfectly align with
reality, the insights they provide are invaluable. They offer a safe and efficient means
for traffic planners and other practitioners to evaluate scenarios that might be chal-
lenging or impossible to test in real life. This high-fidelity virtual representation helps
inform investment plans and policy decisions, equipping practitioners with robust
tools to adjust the Digital Twin according to potential changes in infrastructure or
technology, thereby allowing them to effectively predict and analyze impacts on
system performance.
Urban network design is entering a transformative era with the advent of genera-
tive modeling techniques. This is showcased in the development of a New Green-
field City network. By borrowing network components and the population’s activity
patterns from neighboring cities, planners can now simulate and visualize the
intricate dynamics of a new urban development even before it materializes. This
method leverages Mobiliti to create robust, predictive models that were previously
unattainable.
    The generative design process for a New Greenfield City network is particularly
innovative because it synthesizes not just the physical infrastructure of the network
but also the underlying human activity that drives traffic patterns. This approach
includes borrowing and modifying both geometric elements and Origin–Destination
Time (ODT) data, providing a rich dataset that informs the design and anticipates
the needs of the urban area.
    The effectiveness of this modeling approach is evident in the visualizations that
detail the origins and destinations of all regional trips including a New Greenfield
City network. These visualizations, Figs. 14 and 15, illustrate the density and spread
of trips in the region.
    By adjusting the network elements—links added, removed, or modified—the City
Network Persona evolves. This persona reflects the changes in demand driven by
population activities and commercial vehicle flows, enabling a responsive design
process that adapts to predicted needs.
    Initial model results highlight the existing Bay Area network with vehicle miles
traveled (VMT) before the integration of the New Greenfield City is shown in Fig. 16.
Figure 17 introduces the new city network in the upper right of the network, showing
how it integrates within the existing framework and the accompanying travel demand.
358                                                                 J. F. Macfarlane and I. Babur
Fig. 14 Origins of all trips taken by the population of a New Greenfield City network
Fig. 15 Destinations of all trips taken by the population of a New Greenfield City network
A zoomed-in view, Fig. 18, provides a detailed perspective of how the new network
segments interact with the existing network.
    The final visualization, Fig. 19, demonstrates the regional impact of the New
Greenfield City on VMT. This includes not only the expected increase in VMT
across the entire Bay Area but also specific impacts on highway and county roads
in both the immediate and extended vicinities of Greenfield City. These areas show
significant changes, highlighted in the visual outputs, indicating a substantial increase
in traffic congestion and predicted pressure points. This analysis can be incorporated
as a feedback process for an automated generative design framework that can tune
the Greenfield City network and its associated demand to meet the design objectives
for the Greenfield City, while also addressing regional concerns associated with the
traffic impacts of the new travel demand.
Mobiliti: A Digital Twin for Regional Transportation Network Design …                          359
Fig. 16 Daily vehicle count in the San Francisco Bay area before the addition of the New Greenfield
City network
Fig. 17 Daily vehicle count in the San Francisco Bay area after the addition of the New Greenfield
City network
    The ability to perform such detailed and predictive analyses through Digital Twins
represents a significant leap forward in urban planning. Digital Twins enable plan-
ners to unlock insights that inform investment plans and policy decisions, facil-
itating a proactive approach to urban development. The generative design capa-
bilities facilitated by Mobiliti ensure that planners and stakeholders can explore
various scenarios and their consequences, enhancing understanding and preparation
for real-world implementations. This comprehensive approach not only optimizes the
360                                                               J. F. Macfarlane and I. Babur
Fig. 18 Daily vehicle count in the New Greenfield City network after integration with the San
Francisco Bay area network
Fig. 19 Impact of the integration of the New Greenfield City network with the San Francisco Bay
area network in terms of daily vehicle count
network design but also ensures that new developments, such as the New Greenfield
City, are integrated seamlessly into the larger urban fabric, supporting sustainable
and efficient growth.
Mobiliti: A Digital Twin for Regional Transportation Network Design …               361
Acknowledgements The Mobiliti Platform is the compilation of the work of a lot of excellent
scientists: Cy Chan, Srinath Ravulaparthy, and Max Bremer at Lawrence Berkeley Laboratory, Anu
Kuncheria, Dimitris Vlachogiannis, Colin Laurence, Anthony Patire, Rowland Awadagin Herbert-
Faulkner, and LuLu Lui at University of California Berkeley, Prasanna Balaprakash, and Tanwi
Mallick at Argonne National Laboratory. The work described was sponsored by the US Department
of Energy (DOE) Vehicle Technologies Office (VTO) under the Big Data Solutions for Mobility
Program, an initiative of the Energy Efficient Mobility Systems (EEMS) Program. The following
DOE Office of Energy Efficiency and Renewable Energy (EERE) managers played important roles
in establishing the project concept, advancing implementation, and providing ongoing guidance:
David Anderson and Prasad Gupte.
References
Soumya Singh, Michael Oberle, Daniel Schel, Julian Grimm, Olga Meyer,
and Kai Peter Birke
Abstract The increase in global demand for lithium-ion battery cells has driven the
establishment of numerous manufacturing facilities. However, the complex design
and manufacturing of Li-ion battery cells pose several challenges, such as ensuring
high cell quality while maintaining high process stability and efficiency. During the
usage phase, factors such as operating conditions, and charging/discharging profiles
can significantly impact the performance and longevity of battery cells. Furthermore,
addressing recycling and proper disposal in the end-of-life phase presents substantial
challenges, given the risks associated with hazardous materials and the scarcity of
resources. Hence, effective management of these phases is essential for achieving
sustainable and efficient energy storage systems. Digital twins applied in various
industries have shown promising results in improving product lifecycle management
and smart manufacturing processes. However, it is important to scientifically test
and validate its effectiveness within the specific context of the battery industry. The
main question is whether digital twins can realistically address the growing chal-
lenges of the battery industry, such as degradation evaluation, usage optimization,
manufacturing inconsistencies, second-life application possibilities, etc. Without a
thorough understanding of the benefits of digital twins, it is difficult to claim that they
are the best solution for the battery industry’s challenges. This chapter explores the
potential applications and use cases of digital twins in the battery lifecycle, outlining
the requirements and concepts necessary for implementing digital twins effectively.
This approach ensures a comprehensive understanding of the system’s benefits
and limitations, thereby enabling stakeholders to make informed implementation
decisions.
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2024              363
M. Grieves and E. Hua (eds.), Digital Twins, Simulation, and the Metaverse, Simulation
Foundations, Methods and Applications, https://doi.org/10.1007/978-3-031-69107-2_15
364                                                                          S. Singh et al.
1 Introduction
Lithium-ion batteries (LiBs) are widely favored energy storage systems due to their
promising attributes such as high energy density, low self-discharge property and
long cycle life. The European Battery Alliance foresees a potential market worth
up to 250 billion euros for automotive batteries in Europe from 2025 onwards.
Presently, European production meets only 6% of global LiB demand, prompting
a political target to achieve 30% by 2030 [1]. To support this goal, the Federal
Ministry for Economic Affairs and Energy has allocated nearly three billion euros
for battery cell production, attracting proposals from over 40 German companies [2].
This funding prioritizes sustainability and digitalization, aiming for efficient material
usage, energy conservation, and innovative recycling approaches. In order to meet
such requirements and the rising global demand, the challenges across the battery
lifecycle need to be addressed through cutting-edge technologies [3].
    Driven by the rising market demand for battery electric vehicles (BEVs), consumer
electronics, and battery-driven grid storage, stakeholders across the battery value
chain encounter numerous challenges throughout the battery’s design, manufac-
turing, usage, and end-of-life (EoL) phases. This section will subsequently explore
the challenges within the different lifecycle stages.
    During battery design, challenges can arise from market demands, such as
customers along the supply chain, including original equipment manufacturers
(OEMs), or end-users. Regarding BEVs, users demand high energy density, a long
driving range, long calendar and cycle life, fast and efficient charging, safety, and
low costs [4]. On the other hand, OEMs face design limitations because they aim
for application-specific batteries. The variety of applications potentially influences
the challenges in the design stage [5]. Furthermore, the battery market is dynamic
at the moment, with heterogeneous manufacturers, different packs/cell designs and
formats, electrode designs and formats, different active and inactive material combi-
nations with various levels of readiness, as well as promising future technologies [6,
7]. Bringing all this together, battery design has multiple and fast-changing objec-
tives, conditions, and circumstances, which, as a result, leads to tradeoffs and batteries
that cannot meet all aspects.
    The demands within the battery market generate technical challenges for battery
design and development. It is crucial to reach good performance measures with a
specific battery design. Reaching a high energy density is said to be the overarching
goal [8]. Besides this, safety, thermal stability and temperature ranges, and durability
are further influencing factors.
    The aim of achieving a high energy density and fast charging rates, results in
challenging cooling efforts on the battery system for dissipating and supplying heat.
Furthermore, increasing the size and thickness of individual cells to meet the target
values changes the electrical and thermal conditions at a cell level in a negative
manner [8]. Thus, reducing the amount of inactive material in either a battery cell or
Digital Twin in the Battery Industry                                                 365
a battery system may lead to undesired conditions. In contrast, the system’s thermal
management requirements can be reduced by increasing the cell’s thermal operation
window [5]. In addition to thermal requirements, there is also a tradeoff between
general safety and energy density. In a battery system, there are electrical, thermal,
and mechanical safety protection measures on the cell, module, pack, and vehicle
levels [5]. This results in an effective safety system but ends up in a lower energy
density. Furthermore, battery durability is a critical parameter due to warranty issues
of the battery manufacturer and OEMs. Therefore, in the design phase, selecting
suitable tests and confirming test results for possible future degradation of the battery
during usage under operating conditions is challenging [5].
    Aside from the design challenges due to technical requirements, there are also
issues arising from scarce and expensive raw materials. Since mining and refining
operations are concentrated in a few countries, future resource supply could become
challenging [9]. Efforts are underway to reduce reliance on scarce resources like
cobalt and nickel in battery production [5]. With large-scale battery facilities being
developed, it is crucial to understand how changes in battery design might affect
manufacturing processes and supply chains. Overall, battery design encounters
challenges due to tradeoffs in target values and evolving conditions [7].
    In battery manufacturing, the main objectives are cost savings, quality improve-
ment and fostering sustainability. However, achieving these objectives present chal-
lenges such as scaling up production, enhancing productivity, minimizing scrap, and
optimizing resource and energy efficiency [10]. These challenges can be addressed
at different levels, such as manufacturing operations or process levels.
    As elaborated in [10], certain manufacturing processes have prolonged processing
times, consequently reducing the yield rate. Notably, the electrolyte filling process
stands out as one of the most time-consuming steps. With an increase in the cell’s size,
the reduction of the electrolyte filling process time becomes even more challenging.
Another example related to large cell formats with higher energy density involves
the necessity of adjusting their tab designs according to the specific format, resulting
in the need for greater positioning accuracy and posing handling challenges [8]. Last
but not least, modification in cell designs, as well as the integration of new battery
designs into existing production lines, can be very challenging because they require
significant adjustments to manufacturing processes, machinery, and workflow, often
leading to disruptions, increased costs, and potential delays in production schedules
[7].
    During battery usage, challenges may arise due to user expectations, particularly
concerning the total costs of ownership (TCO), including purchasing costs, electricity
costs, and maintenance costs. To mitigate these expenses, efforts must be made across
the design, manufacturing, and EoL stages, involving process improvements, material
innovations, and implementing circular economy principles [5].
    A major challenge for LiBs is that different operating conditions and load profiles
create individual life cycles that greatly affect the aging of a cell. Consequently,
the aging of each LiB cell within a battery pack differs from one another, leading
to internal inconsistency between the battery cells within one module or pack
[11]. Lastly, challenges such as monitoring performance, ensuring safety, managing
366                                                                          S. Singh et al.
In 2002, the basic concept of DTs was first presented as “The Conceptual Ideal of
PLM” and later in 2005 as the Mirrored Spaces Model by Dr. Michael Grieves. This
Digital Twin in the Battery Industry                                                  367
was the first time that the idea of virtual representation of real objects was applied
in the context of industrial production. Dr. Grieves, who was the Co-director of a
research center at the University of Michigan at the time, is considered a pioneer
who put the concept of DT into practice [14, 15].
    DT technology enables the recreation of real objects or systems in a virtual envi-
ronment. Data is collected from sensors and other sources to create a replica of
the real object. The DT can then be used for various purposes, such as simulating
operating procedures, optimizing processes, or predicting malfunctions.
    The term “Digital Twin” was introduced by John Vickers from NASA in 2010
[15, 16]. It serves as a metaphor illustrating the idea that the DT is a virtual copy
of a real object. Similar to real-life twins, the real object and the DT share certain
characteristics and features. Since then, the number of publications on the subject
has increased enormously [17].
    Jones et al. in [18] recorded a sharp increase in publications on the topic of
DTs as early as 2014. Despite the significant increase in the interest in the DTs in
recent years, research is still at a very early stage, with initial practical applications
remaining tentative and not fully developed. As evidenced by various publications
on DTs in [18, 19], there was already a pressing need to consolidate research and
industry in the early 2020s to establish a common understanding of the topic and
ensure that future research efforts are based on solid foundations for each specific
area.
    The term “Digital Twin” has become ambiguous due to the abundance of scientific
publications and industrial application scenarios. Consequently, organizations have
been formed at the national and international levels with the goal of promoting
and disseminating the concept of DT. However, there exists a range of definitions,
graphic models, and complementary concepts. Nevertheless, the lowest common
denominator in the context of DT remains the Mirrored Spaces Model by Grieves
[20, 21].
    The conceptual process is significantly delayed by the need to define terminology
and concepts for the DT. Consequently, numerous definitions have been broadly used
in both industry and academia for some time now, posing a major challenge for users
to find the right definition and base their work on useful practical specifications. The
lack of consistency and a number of non-consensual solutions, based on a variety
of different definitions and frameworks can potentially risk the benefits that the DT
was originally intended to provide [18].
    For several years now, standardization has been actively pursuing a coordinated
approach to the topic of DTs. The forerunner in this area is the Joint Technical
Committee of ISO and IEC (JTC 1), established to develop global standards for infor-
mation and communications technology (ICT) for business and consumer applica-
tions. Within JTC 1, there is a special subcommittee, SC 41, dedicated to standards for
the Internet of Things. Since the beginning of 2021, this subcommittee has expanded
its scope and terms of reference by establishing a new working group (WG 6) dedi-
cated to DT applications across multiple sectors, including smart manufacturing,
utilities, smart grids, smart buildings, and smart cities [22].
368                                                                                    S. Singh et al.
    It took some time to agree on the definition and formulation of the basic terms,
however, the situation is different in practical research. Here, developers started
a bit earlier and came up with the first solutions. A glimmer of hope is coming
from the field of manufacturing. Due to the advancing digitization in the industrial
environment, the development of concepts and closer practical solutions for the DT
is being tackled more quickly than in other areas.
    One of these areas is battery cell production. In the following sections, we will
take a closer look at the product, the battery cell, or rather the life cycle of battery
cells and the possible applications for DTs in this context.
test and validate their effectiveness in the specific context of the battery industry.
Without a thorough understanding and substantial efforts in quantifying the benefits
of DT, it is challenging to assert that it is a suitable solution to the battery industry’s
challenges [29].
    However, DT implementations present their own set of challenges, such as
ensuring data accuracy and integrity, managing the complexity of integrating various
data sources, addressing cybersecurity risks, and ensuring scalability and interoper-
ability. Additionally, there may be challenges associated with the cost of implemen-
tation, organizational resistance to change, and the need for specialized expertise in
areas such as data analytics and machine learning.
    This brings us to the research questions addressed in this chapter. Firstly, what are
the potential applications of DTs across various stages of the battery lifecycle? We
identify specific use cases where strategic implementation of DTs, as hypothesized,
could alleviate the challenges currently faced. Secondly, how can the development
and implementation of a DT be optimized for a battery-specific industrial context?
By delving into both the theoretical and practical dimensions of these inquiries, we
aim to contribute insights into leveraging DTs for the battery industry.
While it is essential to identify use cases and understand the potential business advan-
tages and functional benefits of DT applications, accurately quantifying these benefits
remains challenging due to the current technological landscape and the ambiguity
surrounding DT classification. Existing literature [3, 27, 30] highlights the theoret-
ical analyses of DT applications in the battery lifecycle; however, translating these
theoretical insights into tangible business advantages necessitates actual implemen-
tation. Recent studies emphasize the need to move beyond theoretical discussions
to quantitatively assess the real-world impacts of DT utilization. In this context,
Fig. 3 summarizes the potential challenges within the battery lifecycle that could be
mitigated through the practical implementation of DT.
    In the following sub-sections, we explore the potential use cases of DTs within
each lifecycle phase to address these challenges.
2.1.1 Objective
With the continuous evolution of battery materials, the pursuit of an optimal battery
remains ongoing, requiring the introduction of new designs and technologies at both
372                                                                        S. Singh et al.
cell and pack levels. This process is both costly and time-intensive. A crucial use case
of DTs is in providing decision-support for battery system design by leveraging real-
world data to reflect the battery’s behavior and performance in various operational
and environmental conditions.
2.1.2 Overview
The trial and error approach in developing new battery designs often results in signif-
icant consumption and wastage of critical raw materials, such as nickel and cobalt.
The motivation behind using a DT during the design phase is to provide stakeholders
with a decision-supporting tool for streamlining the design processes, thereby mini-
mizing material waste and optimizing resource utilization. It aims to facilitate the
exchange of data and information across different lifecycle stages as well as the
prediction of battery behavior and performances using data from various operating
conditions and data from previous product generations, hence, aiding in the design
of more durable and efficient battery system. Additionally, the operational data of
the battery can be used as feedback to adjust the design and optimization processes.
    During the standard design process of LiB systems, several key features need
consideration. For cell design, aspects such as nominal capacity, voltage, internal
resistance, energy and power requirements, cell format (cylindrical, prismatic,
pouch), and electrode material (e.g., lithium cobalt oxide, lithium iron phosphate)
play a crucial role. Safety features, including prevention of overcharging, discharging,
and thermal runaway, along with cooling systems and electronic control through
Battery Management System (BMS), are vital considerations [31].
    Recent studies [32, 33] have proposed various DT frameworks for smart product
design. Applying DT in the product design stages allows for providing a quantita-
tive design tool for efficient and optimal design decisions using data from previous
product generations [34], as well as operational data from the usage phase. In the
design process, battery models, such as electrochemical-thermal models or other
simulation-based and data-driven methods, play a crucial role in scenario simula-
tion. These tools are instrumental for enhancing battery design, reducing long design
iterations, and expediting the time-to-market.
This use case falls under the category of a DTP of the product. According to Fig. 1,
there is a physical component and a cyber-component of the product DTP. Aligning
the use cases according to the classification system in Fig. 1 helps establish the
system boundary for implementation. This process of defining the system boundary
will be applied to all subsequent use cases outlined in the following sections.
    The battery DTP primarily comprises a combination of data, models, and algo-
rithms already employed for battery simulations. In the design phase, the battery
Digital Twin in the Battery Industry                                                  373
DTP can determine the impact of various parameters on system behavior, identify
problems within the system, and simulate deterioration deterrence options [35].
    Figure 4 depicts the DTP for battery design. On the left side of the figure, each life
cycle phase is associated with the types of DTs based on their definitions (in Sect. 1.2),
namely DTP, DTI, and DTA. The battery DTP streamlines the design processes by
utilizing the design data, experimental data, real-time and environmental-impacted
data to feed the battery models. Various modeling approaches are available for simu-
lating the battery states (such as state of health, state of charge, etc.), performance,
aging and safety of LiBs. These include electrochemical models (like pseudo-2D-
P2D model, single particle model SPM, etc.), electrical models (equivalent circuit
model ECM), thermal models, mechanical models, empirical models, data-driven
approaches, and hybrid approaches. These models are coupled with algorithms, listed
in Fig. 4, to streamline the scenario simulations.
    The DTP serves as a virtual prototype, simulating design scenarios and facilitating
efficient testing and redesigning. This reduces costs by identifying flaws before actual
production. By harnessing the capabilities of the DTP, battery designers can create
374                                                                         S. Singh et al.
dynamic simulation scenarios, optimize auxiliary systems, and accurately predict the
performance of physical products, enhancing design efficiency and product quality.
   Let’s consider a scenario where an existing battery pack is currently utilized in
EVs. Now, there arises a new imperative for EV manufacturers: the design of a novel
battery pack featuring a new cell chemistry that promises higher energy density.
Notably, all other aspects of the pack will remain unaltered. In light of this, we
explore the use case of a battery DTP within this context. At the model level, changes
in cell chemistry affect a wide range of electrochemical parameters, such as particle
radius, electrode thickness, conductivity, and diffusivity of the active materials. These
affected parameters can be provided by the material supplier, measured through test
experiments, or identified through various parameter identification algorithms. The
existing pack design also has DTI and DTA from the manufacturing, usage, and EoL
phases. Using the information on the new design requirements and the DTI/DTAs of
the previous product version, a DTP of the new product prototype is developed.
   The DTP of the new design offers a transformative solution for battery designers
by integrating data from various lifecycle phases, facilitating simulation, testing,
and optimization of battery performance in a virtual environment. Consequently,
before initiating the entire value chain, the compatibility of the new design at both
the module/pack and system levels can be tested virtually. When designing or opti-
mizing the current battery design to align with operational needs, the real-time data
provided by the DTIs offer insights into the battery pack’s operating cycle, working
environment, and other critical information not easily obtained through traditional
laboratory experiments. As a result, the DTP promotes battery design, development,
and research, ushering in a more efficient and informed approach to advancing battery
technology.
Digital Twin in the Battery Industry                                                   375
2.1.4 Challenges
. The models used for simulation the design scenarios are the core element of the
  DTP. Ideally, the battery model is a high-fidelity multiscale that can incorporate
  both temporal and spatial dynamic features. One of the challenges is to quantify
  and manage the uncertainty inherent in these scenario simulations and subse-
  quence choices based on them. Furthermore, the verification of such models is
  complex and computationally intensive.
. The challenge associated with the effective integration of data from various life
  cycle phases and technical domains during the implementation of a DTP for
  battery design optimization lies in ensuring seamless communication and compat-
  ibility between disparate datasets. Achieving this integration requires overcoming
  technical barriers, standardizing data formats, addressing interoperability issues,
  and ensuring data consistency and accuracy across different sources.
. The aggregation of multiple datasets used to feed the models is associated with
  multidisciplinary challenges concerning the sharing methods, privacy, security,
  and governance systems.
2.2.1 Objective
Batteries are required to adhere to strict performance and quality standards. However,
manufacturing processes include non-deterministic factors like material inconsis-
tencies, environmental conditions, human error, and equipment wear. These factors
contribute to variability in the outcomes of the manufacturing process. The implemen-
tation of DTs in battery manufacturing is aimed at mitigating these inconsistencies in
order to minimize scrap, increase throughput, and enhance reliability of the battery
cell during use.
2.2.2 Overview
To achieve the mentioned objectives, different types of DTs are required, each serving
specific purposes. If the aim is to reduce scrap and ensure cell reliability, the DT of the
process becomes paramount. Conversely, if the objective is to increase throughput,
376                                                                       S. Singh et al.
then the DT of the resource is key. Additionally, the DTI of the product also plays a
crucial role in supporting both objectives.
    While the DTP of the product is critical during product design and provides the
baseline structure and data for the DTI, the DTI of the product becomes key during
the manufacturing life cycle. There exists an essential link between the DTI of the
product and the DTI of the process and resource. On one hand, it collects data
throughout the entire manufacturing process, while on the other, it identifies if the
asset is within specifications. DTAs of the product enable statistical analysis and
machine learning approaches to derive models and rules that can be leveraged for
the creation of DTP of the processes.
    In a use case aimed at ensuring cell quality by maintaining manufacturing process
outcomes consistently within specifications, the DTI of the process utilizes informa-
tion from the product DTI regarding process results and product specifications for
its calculations. For example, this can allow for the dynamic calculation of a recipe
for adjusting the dryer during coating based on the viscosity of the slurry. Similarly,
the DTI of the product supports the DTI of resources in the battery manufacturing
process. Information on the result of each process step is essential information for
the DTI model in terms of calculating the wear level. Based on this information, the
DTI can determine when maintenance procedures need to take place to minimize
resource downtime (Fig. 5).
Considering the DTSM, both the DTI of the product and the DTI of the process and
resource have descriptive and executive components. Figure 6 shows a framework for
Digital Twin in the Battery Industry                                               377
DTs in battery manufacturing for PPR and illustrates their interrelations, including
the primary data required.
    For the DTI of the product, quality data is collected at each transformative stage
of the manufacturing process. This data is continuously enriched from the mixing of
slurry to the manufacturing of electrodes, cell assembly, and finally, cell formation.
The collection of quality data is particularly crucial during manufacturing, not only
for compliance with customer standards but also for internal processes.
    The product’s DTI includes an executive component for evaluating the current
state of its quality, employing either statistical or machine learning models. Alterna-
tively, it can utilize the DTP and its models developed during battery design. Quality
data collected up to the current work in progress is utilized for this purpose. For
example, in slurry mixing, quality data encompasses viscosity and homogeneity.
Quality models can assess whether these values fall within acceptable parameters
and predict their impact on future manufacturing steps. Typically, these models are
constructed using machine learning [36, 37].
    On the manufacturing shop floor, each process has a separate DTI due to its
distinct nature. For example, the mixing and coating processes are continual, whereas
assembly and formation are discrete. It is evident that each manufacturing step can
differ significantly from another, even without considering the specific techniques
involved. The process DTI comprises descriptive and executive components. The
descriptive parts include all parameters (e.g., time or temperature for mixing) that
define the process. The executive part models the process behavior with the aim of
providing actionable information for process control. Typically, this involves a recipe
that includes process parameters for the current run of the process.
    Information of the product DTI is crucial for the process DTIs. On one hand,
it facilitates adapting the process to the current product by retrieving the current
characteristics achieved by preceding processes. On the other hand, it can provide
indications of effectiveness by retrieving the outcomes of subsequent processes to
adapt for future process runs [38]. These dynamic short-term adjustments enabled
by the DTAs of the product allow for more fundamental process adjustments by
conducting deeper analyses on hundreds, if not millions, of occurrences [39].
    The resource DTI functions similarly to the process DTI in interactions and
capabilities but focuses on machine data rather than process data. For instance,
in a battery manufacturing sub-process like tab welding, executed using different
378                                                                        S. Singh et al.
machines such as laser and resistance welders, the process DTIs control the process,
while resource DTIs handle energy consumption and machine behavior. Models
derived from resource DTIs optimize machine-related factors such as maintenance
scheduling or cycle times, employing techniques such as discrete event simulation
[40] or machine learning models [41].
2.2.4 Challenges
. As described in the previous paragraphs, the interplay of the PPR DTIs plays
  a crucial role. However, besides communicating with each other, this requires
  gathering data from various systems, such as ERP, MES, machines, and tools.
  Integrating this data from diverse sources and ensuring its accuracy and timeliness
  is a significant challenge. This is exacerbated by the vast amount of data that
  needs to be aggregated. Considering that several hundred thousand battery cells
  are produced each day, the data volume that needs to be processed and stored is
  immense. Hence, careful planning of the system architecture is required.
. In addition, the creation and maintenance of the DTIs for process and resource,
  particularly the executive models, can be challenging. For both, accurately
  modeling the dynamic behavior of the real asset is essential. This in particularly
  challenging in battery manufacturing processes, which are complex to model. A
  prime example is the coating process. While the fundamental steps required for
  coating are well understood, modeling becomes difficult due to numerous influ-
  encing factors (e.g., slot control, dryer control) that affect the behavior of the
  model. Furthermore, the relationship between processing conditions (e.g., slurry
  composition, coating speed, drying temperature) and the resulting microstructure
  is an area of active research.
    Addressing these challenges requires strategic integration and a robust system
architecture. One potential approach is to implement standardized description
formats and interfaces that enable uniform models and access to diverse data sources.
This approach could also facilitate abstracting physical data storage, separating repre-
sentation from big data high architecture design. Furthermore, the creation and
maintenance of executive models could be streamlined by integrating simulation
models and machine techniques. This integration can help mitigate situations where
a lack of data impedes machine learning efforts, or validate data-driven models with
simulations.
Digital Twin in the Battery Industry                                               379
2.3.1 Objective
The objective of a DT during the battery usage phase is to predict and investigate the
degradation mechanisms and potential failures, thereby optimizing performance and
extending the service life. Through the integration of battery operation data with the
models, a DT enables proactive management of battery health and performance.
2.3.2 Overview
This use case falls under the category of DTI of a product, specifically the battery
system. According to Fig. 1, there are both physical and cyber components of the
product DTI. In the context of this use case, a DTI represents a collection of data
generated when the battery system is under operation in a certain application, such
as the data from the BMS, service information, serial number, etc.
    The DTI of the individual cells, modules, and battery systems are already created
during the manufacturing phase. Battery systems are complex products composed
of multiple modules and a significant number of battery cells. The DTI of a battery
system is composed of the DTAs of the cells and the DTAs of the modules. During
the usage phase, the DTI of the battery system is connected to the specific physical
twin and remains connected throughout the entire life of that physical twin (i.e., the
battery system).
    Figure 6 represents a battery system DTI. The behavioral profile of the battery
is recorded as operational data during its actual usage. This typically includes data
acquired from on-board BMSs, consisting of information from current, voltage, and
temperature sensors. In most cases, BMS data is stored locally. However, recent
research efforts have focused on mapping real-time battery data from BMS to cloud-
enabled systems. This approach can potentially minimize local computational needs
and aggregate large datasets to improve the performance of ML-based algorithms.
    The DTI can continue to utilize the models and algorithms in the DTP created
during the design phase. By feeding operational data into the DTP developed during
the design phase, the battery model needs to evolve to align with battery usage
profiles. This ability for model evolution is a vital component of the algorithm,
providing the battery DTI with essential “intelligence” during the usage phase.
The exploration of model evolution represents an intriguing research gap within
battery modeling and is the focus of extensive investigation. This concept is further
elaborated upon and applied in a case study presented in Sect. 3.
    The DTI’s result for this use case involves degradation assessment in terms of SOH
estimation and RUL forecast, along with optimization measures based on predicting
voltage response. These capabilities enable the derivation of maintenance measures
during servicing, facilitating efficient troubleshooting, improving customer satisfac-
tion, and providing advanced monitoring capabilities. While individual users may
not delve deeply into battery degradation mechanisms, organizations utilizing battery
systems for applications such as EVs, smart grids, or industrial power require a
comprehensive understanding of how the usage patterns influence system aging.
These users benefit from the results of a battery DTI in terms of optimizing charging
patterns, accurate range estimation, estimating system reliability, and responsible
energy management.
Digital Twin in the Battery Industry                                                 381
2.3.4 Challenges
    To address these challenges, standardizing reference DTIs at both the cell and
module levels is crucial. This will simplify the development and maintenance of
the architecture. Additionally, progress in Li-ion battery modeling techniques, along
with the establishment of standardized data integration methods, and further research
in model evolution methods are essential steps. These endeavors will enhance archi-
tecture development, improve model accuracy, streamline data integration, and refine
battery models, ultimately optimizing system performance and fostering innovation
in the energy storage sector.
2.4.1 Objective
2.4.2 Overview
With the increase in demand for LiBs, prolonging the overall utility of the batteries
is becoming increasingly important. Due to the presence of critical materials, such
as cobalt and nickel, and the high manufacturing costs of battery packs, there is
economic, political, and environmental interest in utilizing existing battery system
in second-life applications. Currently, there is insufficient information available for
choosing a second-life application. Examples of second-life applications include
382                                                                        S. Singh et al.
power grid storage systems, automated guided vehicles like forklifts, and home
storage systems for storing photovoltaic electricity temporarily. In such applications,
the battery system experiences lower demands in terms of external temperatures,
charge/discharge currents, and SOH.
   DTs provide an approach to tackle this challenge by creating a digital representa-
tion of the battery´s past life, including design details, manufacturing details, usage
history, and environmental interactions. This provides a comprehensive insight into
both the structural and functional aspects of the batteries. Such information is crucial
for designing, planning, and executing the processes of remanufacturing, recycling,
and repurposing. Additionally, since the disassembly of batteries is safety–critical,
this information also enables process automation.
   The concept of a DT at battery EoL aligns with that of a Digital Battery Passport
(DBP). According to the European Commission legislation, all industrial batteries
and EVBs should implement a DBP by 2026 [43]. A DBP, as defined by the Euro-
pean Commission, is an electronic record (e.g., QR code) of an individual battery
containing crucial information about the sustainability characteristics of batteries,
such as capacity, performance, durability, and chemical composition. It aims to
support the scaling of circular economy strategies such as disassembly, repurposing,
remanufacturing, and recycling.
This use case falls under the category of DTI/DTA of a product, specifically the
battery system. When data from a population of battery systems is collected and
aggregated, it is referred to as a DTA. DTI/DTAs provide a method to aggregate the
entire lifecycle data of battery systems.
   As illustrated in Fig. 7, data relevant to EoL decision-support is collected from all
other lifecycle phases. Design information from the DTP provides insights into the
original specifications and performance characteristics of the battery. Operational
data from the battery DTI/DTA is used to estimate SOH and predict RUL of the
battery in a second-life application. Promising methods exist for classifying end-
of-life batteries using physics-based models or data-driven models and algorithms.
Manufacturing data from the DTI offers details about the assembly format, materials,
and quality control processes used in producing the battery. The design-support tool
can use such information to identify inherent limitations or weaknesses in the battery
design that may impact its performance or safety during a second life.
   The DTI/DTA at EoL facilitates the decision-making process to identify the most
suitable second-life applications. Similarly, the data can also be used to determine
other suitable pathways such as remanufacturing, repurposing, or recycling.
   We propose the integration of DTIs as a solution for the DBP, providing a readily
available implementation. The technical design and operation of a DBP involve struc-
tured product-related data with a predefined scope, agreed data ownership, and access
rights conveyed through a unique identifier. By linking DTIs with DBP, stakeholders
Digital Twin in the Battery Industry                                               383
Fig. 7 DT at EoL
can access a wealth of information regarding the origin, usage, and disposal consid-
erations of batteries, thereby enhancing decision-making processes and promoting
sustainable practices.
   DTIs contribute to a closed-loop lifecycle for batteries, supporting a circular
economy and responsible resource management. Their application promises signif-
icant environmental and economic benefits, transforming battery lifecycle manage-
ment toward sustainability and efficiency.
2.4.4 Challenges
issues related to data silos, lack of interoperability, reliability concerns, and clarity
regarding responsibilities at various EoL stages. Ultimately, standardized reference
models play a crucial role in streamlining.
of engineering data as part of AAS. They present the AAS completion method based
on the AutomationML standard to facilitate the capture of the AAS engineering
dataset and its export to an AAS serialization.
    The AAS can be considered from two perspectives: the virtual representation as
an asset description and the technical functionality as a smart manufacturing service.
In order to enable the effective provision of information for DT applications, Park
et al. [50] propose Virtual REpresentation for a DIgital twin application (VREDI),
which is an asset description for the operational procedures of a work-center-level
DT application.
    These descriptions are aligned with the intended use of a DT, reflecting the
perspectives of involved professionals such as process designers, control special-
ists, and managers. Cimino et al. [51] introduce a paradigm, the Digital Multiverse,
to encompass major DT interpretations, not only in terms of data integration but also
by establishing and enforcing consistency rules involving both data and models.
    As defined in the Reference Architectural Model for Industry 4.0 (RAMI 4.0), an
AAS represents a practical embodiment of the DT concept. Its realization involves
the integration of operational technologies and information and communication tech-
nologies. Ye et al. [52] aim to present the current status of AAS development, design
an intuitive method for implementing AASs, and develop an AAS-enabled digital
solution for cyber-physical applications in the manufacturing sector.
    The survey by Jacoby et al. [53] of open-source implementations of the AAS
involved a structured search on GitHub and Google Scholar, which led to the selection
of four implementations for in-depth analysis. This paper represents a first attempt to
provide a comprehensive comparison of AAS implementations: ASX Server, Eclipse
BaSyx, FA3ST Service and NOVAAS.
    In conclusion, it can be summarized that the AAS is now gaining traction after
many years of development and testing. Additionally, the AAS is open to integrating
existing standards and has the flexibility to cover all life cycle phases of an asset.
It can also be extended to design DTs for various application areas and domains.
Furthermore, a growing open-source community is also contributing to its widespread
adoption.
tangible or intangible. References to other AAS can also be managed within an AAS
in order to describe certain relationships or dependencies. Due to this nature of the
AAS, the DT types – DTP, DTI, DTA – can be implemented with the help of the
AAS.
   The two dimensions of the DTSM, Lifecycle and PPR Model, can also be imple-
mented based on these characteristics of the AAS. Part of the third dimension of the
DTMS is already represented by the relationship between the AAS and the asset.
For the remaining cyber part (Descriptive and Executive), a deeper consideration of
the AAS metamodel is necessary. The AAS describes the aspects of a specific DT
of an asset using sub-models. These sub-models represent the semantic description
for such a DT aspect and consist of a collection of properties that describe certain
characteristics of the asset, such as the length, height, and width of a product or the
current temperature. These properties form the descriptive part. The executive part
consists of operations that can be defined as part of a sub-model. Operations take
properties as input parameters and return properties as results. In this way, executable
functions and services can be described and called.
   The actual execution of an operation is implementation specific; for instance, a
simple arithmetic operation or a simulation model can be executed.
   For this reason, the AAS is also ideal for implementing a DT for the battery life-
cycle, as discussed in more detail in the following sections through implementation
examples.
Building upon the understanding of the AAS, we will now examine how it serves
as an ideal framework for implementing a battery DT during the operation phase to
estimate the state of the battery. In this case study, a battery DT software application
is implemented, which can be installed and accessed by various stakeholders, such
as battery specialists seeking insights into the cell behavior in actual applications
during usage. Models and battery data are associated with the assets, thereby serving
as the backend of the software application, which is then used for practical use cases.
    There is no one-size-fits-all approach for implementing this software application
because its usage heavily depends on the IT infrastructure of the application land-
scape. Therefore, a tailored implementation strategy is necessary to accommodate
the unique characteristics and requirements of each application environment.
    Drawing inspiration from the foundational principles outlined in [54–56], a
DTI for state estimation of LiBs is implemented, and the architecture of the
implementation is presented in Fig. 8.
    The DTI components are as follows:
1. Asset: This refers to the real-world entity or system that the DT aims to replicate.
   For this implementation, the asset is a battery pack. Please note that an asset of
Digital Twin in the Battery Industry                                                387
    a cell or module can be referenced to the AAS of the battery pack. This also
    implies that the DTI of a battery pack is a DTA of the battery modules.
   The AAS, implemented through the AASX package explorer, is shown in Fig. 9. It
consists of a number of sub-models in which all the information and functionalities of
the battery – including its features, characteristics, properties, statuses, parameters,
measurement data, and capabilities are stored. The granularity of operation data fed
into the DTI depends on the interface responsible for real-time data transfer from the
BMS.
   By creating an AAS for the battery pack, we have established a standardized
framework for accessing, managing, and exchanging data related to the battery’s
operation, maintenance, and performance. This AAS enables seamless integration
of operation data with the models, facilitating holistic monitoring, analysis, and
optimization of the battery pack’s usage and health. Through the AAS, stakeholders
can gain valuable insights into the battery’s state, make informed decisions, and drive
continuous improvement in its performance and reliability.
2. Data Collection and Device Control Entity: The AAS refers to this entity, which
   comprises two sub-entities. The first sub-entity is responsible for collecting data
   from different sources. Battery-related data such as design data, experimental
   data, operation data, data derived from simulations, and expert knowledge about
   battery behavior form the basis data for battery DT. These data are used for model
   development, verification, evolution, and execution. It is important to note that
388                                                                      S. Singh et al.
   the methods/interfaces used for acquiring and storing these diverse data sets may
   vary. The second sub-entity is the control unit, like the BMS, which sends control
   programs to the battery system when adjustments are needed.
3. Core entity: The core entity acts as the “brain” of the DTI, serving as a cognitive
   center. The functions performed by the core entity include data management,
   data analytics, model development and management, model verification, result
   generation, and model evolution. Additionally, the interfaces that connect the
   software components are also part of the core entity.
   The data management module processes and manages all types of data acquired
in the data collection entity. In the model management module, all models can
be stored, used, managed, and updated. The DT battery models include physics-
based electrochemical models, ageing models, data-driven model lifecycle prediction
models, etc. The algorithms responsible for generating results (estimations, predic-
tions, or forecasts, and battery behavior evaluation) from the models or the data are
also integrated in this module. As the battery is used the DT accumulates data from
the usage phase such as charging/discharging profiles, working environments sensor
acquisitions, and maintenance data. By combining historical and real-time data, the
DTI parameters can be updated and evolved using methods such as neural networks,
Bayesian regression, maximum likelihood parameter estimation, and other param-
eter estimation algorithms. This model evolution module works continuously to
enhance the accuracy of the DTI model.
4. User Entity: The DT software application is a web service accessed by the user.
   This application should include some type of visualization module for users to
   access the results of the DTI. For this implementation, we utilized the AAS Web
   GUI [57]. AAS Web GUI is a user interface provided by Basyx for interacting
   with the AAS Server and AAS Registry through a web browser. For further details
Digital Twin in the Battery Industry                                              389
    on asset integration through Basyx, readers are referred to [58]. Figure 10 shows
    a sample view of the AAS GUI for the battery pack.
   The end user of the DTI is not necessarily the end user of the battery system as
well. The DT developer is responsible for accessing the core entity and its connection
with the other entities.
   Implementing a DT for state estimation of LiBs cells holds significant promise
in enhancing our understanding of its degradation mechanisms and in optimizing its
usage in applications. The AAS framework enables an interface for bi-directional
data exchange between the physical and the virtual space. However, a limitation
remains that the accuracy of the DT heavily relies on the quality and granularity
of data fed into the battery models. This challenge can be mitigated by employing
reduced-order and adaptive models, which offer possible ways to improve the model’s
robustness. Despite these challenges, a DT implementation represents a significant
step forward in advancing battery management practices. By continually refining
methodologies and overcoming limitations, we can unlock the full potential of DTs
in revolutionizing the way we monitor, analyze, and optimize LiBs for improved
performance and reliability in real-world settings.
4 Conclusion
Across the sections of this chapter, we discussed the use cases in different lifecycle
phases of a battery, illustrating the implementation framework and the corresponding
challenges encountered during implementation. We shared insights into the prac-
tical considerations and strategies necessary for the implementation. Moreover, we
described an example of our own implementation, offering a tangible demonstration
390                                                                                      S. Singh et al.
of how the theoretical concepts discussed throughout the chapter can in applied in a
in real-world battery applications.
    In conclusion, this chapter highlights the transformative potential of DTs in revo-
lutionizing decision-making and operations throughout the battery lifecycle. From
real-time monitoring to predictive maintenance and continuous data-driven insights,
digital twins emerge as powerful catalysts for efficiency, innovation, and sustain-
ability in the dynamic landscape of battery technology. The chapter underscores the
profound impact of the DT concept on advancing capabilities and environmental
considerations associated with batteries across various applications, setting the stage
for a paradigm shift in how we perceive and interact with battery systems.
References