Digital Twin AI Opportunities and Challenges from Large Language Models to World Models
đ Original Paper Info
- Title: Digital Twin AI Opportunities and Challenges from Large Language Models to World Models- ArXiv ID: 2601.01321
- Date: 2026-01-04
- Authors: Rong Zhou, Dongping Chen, Zihan Jia, Yao Su, Yixin Liu, Yiwen Lu, Dongwei Shi, Yue Huang, Tianyang Xu, Yi Pan, Xinliang Li, Yohannes Abate, Qingyu Chen, Zhengzhong Tu, Yu Yang, Yu Zhang, Qingsong Wen, Gengchen Mai, Sunyang Fu, Jiachen Li, Xuyu Wang, Ziran Wang, Jing Huang, Tianming Liu, Yong Chen, Lichao Sun, Lifang He
đ Abstract
Digital twins, as precise digital representations of physical systems, have evolved from passive simulation tools into intelligent and autonomous entities through the integration of artificial intelligence technologies. This paper presents a unified four-stage framework that systematically characterizes AI integration across the digital twin lifecycle, spanning modeling, mirroring, intervention, and autonomous management. By synthesizing existing technologies and practices, we distill a unified four-stage framework that systematically characterizes how AI methodologies are embedded across the digital twin lifecycle: (1) modeling the physical twin through physics-based and physics-informed AI approaches, (2) mirroring the physical system into a digital twin with real-time synchronization, (3) intervening in the physical twin through predictive modeling, anomaly detection, and optimization strategies, and (4) achieving autonomous management through large language models, foundation models, and intelligent agents. We analyze the synergy between physics-based modeling and data-driven learning, highlighting the shift from traditional numerical solvers to physics-informed and foundation models for physical systems. Furthermore, we examine how generative AI technologies, including large language models and generative world models, transform digital twins into proactive and self-improving cognitive systems capable of reasoning, communication, and creative scenario generation. Through a cross-domain review spanning eleven application domains, including healthcare, aerospace, smart manufacturing, robotics, and smart cities, we identify common challenges related to scalability, explainability, and trustworthiness, and outline directions for responsible AI-driven digital twin systems.đĄ Summary & Analysis
1. **Concept and Integration of Digital Twins with AI**: - Digital twins are precise replicas of physical systems, enhanced by AI integration. - For example, automotive manufacturers can significantly reduce development time and costs using digital twins.-
Modeling Physical Systems:
- Physics-based modeling is essential for understanding complex systems but becomes even more effective when combined with AI.
- Digital twins synchronize with real-world systems to analyze and predict in real-time.
-
Impact of AI on Digital Twins:
- AI technologies enable automation and optimization of digital twins.
- Large Language Models (LLMs) and generative AI transform digital twins from simple simulation tools into advanced cognitive systems capable of reasoning and imagination.
đ Full Paper Content (ArXiv Source)
1.1
Introduction
“What I canât create, I donât understand.”
â Richard Feynman
Digital twins (DT), as precise digital representations of physical twins (real-world entities or systems), are meticulously designed to maintain a bidirectional connection with their real-world systems, enabling state synchronization for monitoring, prediction, optimization, and decision support . Beyond mere replication, digital twins embody a paradigm shift from static digital mirrors to dynamic, continuously learning reflections of reality. Due to the inherent advantages of predictive analytics, dynamic system simulation, and operational optimization that digital twins provide, this technology has been widely used for healthcare , biological domain , urban planning and management , manufacture , and science . As NVIDIAâs founder and CEO, Jensen Huang, stated in a keynote at the Berlin Summit for the Earth Virtualization Engines initiative, AI and accelerated computing will revolutionize our understanding of complex systems , highlighting a new era where digital twins evolve from analytical tools into intelligent agents that learn, predict, and act upon the physical world.
Technically, a digital twin operates by integrating comprehensive sensor data from a studied object, such as a wind turbine, where sensors are strategically placed to monitor crucial performance metrics including energy output, temperature, and weather conditions. This data is continuously transmitted to a processing system, which applies it to a virtual replica of the physical object. Utilizing this up-to-date digital model, various simulations can be conducted to analyze performance issues and devise potential improvements. The ultimate goal of this process is to extract insightful knowledge from the simulations, which can then be applied to enhance the real-world object, optimizing its efficiency and functionality. A more familiar example of a digital twin is Google Map, which fuses satellite imagery, GPS data, and real-time traffic to maintain a constantly updated mirror of the physical world  . This continuous feedback loop between sensing, modeling, and adaptation forms the conceptual foundation of AI-driven digital twinsâsystems that not only reflect the world but also learn from it to guide real-world actions.
Over the years, the concept of digital twins has evolved significantly, increasingly integrating with artificial intelligence (AI) breakthroughs to transform how we simulate and predict the behaviors of physical systems . Digital twins and machine learning (ML) are closely intertwined, enhancing predictive maintenance and decision-making across various industries. Early systems relied on traditional machine learning algorithms to support predictive maintenance and fault detection . However, as data volumes and system complexity expanded, deep learning (DL) emerged as the cognitive core of digital twins, empowering them to extract intricate spatiotemporal patterns and emulate complex dynamics. Architectures such as convolutional , recurrent , and graph neural networks have enabled digital twins to move from modeling observed behavior to reasoning about unobserved mechanisms. This integration marks a fundamental transition: AI is no longer merely a component within digital twins but the intelligence that animates them.
/>
With the advent of large-scale AI models and foundation architectures, the synergy between AI and digital twins has entered an unprecedented phase . Recent breakthroughs, from LLM-based autonomous agents to world models , demonstrate how AI can emulate, reason, and even imagine complex physical systems. For instance, a multi-agent system framework that applies an LLM is utilized to automate the parametrization of process simulations in digital twins . Moreover, NVIDIA Cosmos provides a world foundation model that generates photorealistic synthetic environments for digital-twin simulations in robotics and autonomous systems .
The convergence of AI technology and digital twins promises to not only enhance the fidelity and responsiveness of these virtual models but also to redefine the boundary between simulation and intelligence. Augmented by adaptive learning and generative reasoning, AI-driven digital twins can anticipate faults before they occur, personalize interventions, and autonomously manage complex systems. Such capabilities herald a future where digital twins evolve into trustworthy, explainable, and human-aligned partners in science, industry, and healthcare. .
Given this transformation, there is an urgent need to consolidate knowledge across the rapidly diversifying landscape of AI-powered digital twins. This paper provides a comprehensive, AI-centered overview of digital twin technologies. We begin by tracing the history of digital twins to establish the conceptual foundation. We then present a four-stage lifecycle that organizes how AI empowers digital twins: modeling the physical twin through physics-based methods and data integration, mirroring it into executable simulators, intervening through prediction, anomaly detection, and optimization, and ultimately achieving autonomous management via large language models and intelligent agents. Given the diverse applications of digital twins, we will examine how AI technology enhances their implementation across different domains such as healthcare, biological systems, and industry. Lastly, we will discuss the existing challenges and issues in using AI technology for digital twins and offer insightful recommendations for future research directions.
Major Contributions
To the best of our knowledge, this paper provides the AI-centered conceptual synthesis of digital twins as evolving intelligent systems. Unlike prior domain-specific reviews, we present a unified framework that connects the physical, digital, and cognitive layers of this emerging paradigm. The contributions of this paper are summarized as follows.
-
We conceptualize digital twins as evolving AI systems, distilling a four-stage lifecycle: describing the physical twin, mirroring the physical twin to digital twin, intervening in the physical twin, and autonomously managing the physical twin. This layered perspective reveals how AI continuously enhances the fidelity, intelligence, and autonomy of digital twins.
-
We provide an in-depth analysis of the integration between physics-based modeling and data-driven learning, highlighting the transition from traditional numerical methods to physics-informed neural networks, neural operators, and foundation models for physical systems. This synthesis clarifies how physical principles and learning algorithms can jointly improve interpretability, generalization, and reliability in digital twin modeling.
-
We analyze the rapid development of generative AI, including large language models, diffusion models, and world simulators, and examine their role in enabling reasoning, communication, and imagination within digital twins. These technologies transform digital twins from passive simulation tools into proactive, self-improving cognitive systems capable of autonomous understanding and creative scenario generation.
-
Through extensive review across eleven application domains, we identify both universal challenges such as scalability, explainability, and trustworthiness, as well as domain-specific requirements in areas including healthcare, aerospace, energy, and education. These observations reveal how AI-driven digital twins are evolving toward more intelligent, interoperable, and ethically responsible ecosystems, highlighting key directions for future exploration and interdisciplinary collaboration.
Organization
To guide readers through this interdisciplinary synthesis, this paper is organized into seven main sections. Section 2, History of Digital Twins, reviews the conceptual origins and technological evolution of digital twin systems. Sections 3â6 form the methodological core of this work, presenting a progressive framework that models, mirrors, intervenes, and autonomously manages the physical world through its digital counterpart. Specifically, Section 3, Modeling the Physical Twin, describes how physical systems are represented through physics-based and data-driven modeling. Section 4, Mirroring the Physical Twin into the Digital Twin, explains how these models are instantiated and visualized within virtual simulators. Section 5, Intervening in the Physical Twin via the Digital Twin, focuses on predictive modeling, anomaly detection, and optimization techniques that enable human-in-the-loop decision-making. Section 6, Towards Autonomous Management of the Digital Twin, advances this paradigm toward AI-driven autonomy, highlighting large language models, foundation models, and intelligent agents as enablers of self-managing digital twins. Finally, Section 7, Applications, demonstrates how these methodological principles are applied across diverse domains such as healthcare, aerospace, smart manufacturing, and robotics, illustrating the broad impact of digital twin technologies in real-world systems. Finally, Section 8, Open Challenges and Future Directions, discusses the key open problems and outlines future research directions for building scalable, trustworthy, and autonomous digital twin systems.
History of Digital Twins
The concept of the digital twin was formally introduced in 2002 by Michael Grieves during a presentation at the University of Michigan. This presentation emphasized the establishment of a product lifecycle management center that integrated both real and virtual spaces, along with data flows to enhance efficiency and innovation in product development and management . While the terminology around digital twins has evolved over the years, the fundamental idea of merging digital and physical twins has remained constant.
Interestingly, the practice of digital twinning dates back to the 1960s, long before the term was coined. NASA was among the early adopters, using basic forms of digital twins for space missions. One notable example is the Apollo 13 mission, where simulations using digital twin concepts played a critical role in bringing the crew safely back to Earth . These early applications demonstrated the potential of digital twins in enhancing design, maintenance, and operational efficiency across various industries.
Companies like Rolls-Royce have been pioneers in adopting digital twin technology. They have utilized digital twins to customize repair processes for engine parts, automating and optimizing maintenance practices based on the specific geometries of these components . Similarly, in the aerospace industry, Boeing employed digital twins for the 787 Dreamlinerâs battery systems to improve safety and manage risks more effectively . Airbus has also embraced this technology, using digital twins for their A350 XWB aircraft to enable real-time performance monitoring, which has led to substantial improvements in fuel efficiency and reductions in emissions .
Recent advancements in digital twin technology have continued to drive significant innovations across multiple industries. In manufacturing, healthcare, construction, automotive, and urban planning, digital twins are becoming indispensable tools. For instance, Tesla has harnessed digital twins to accelerate vehicle development, while the Mayo Clinic has used them to advance personalized medicine . Additionally, architects and city planners are leveraging digital twins to enhance project management and urban development, integrating AI and IoT to optimize resource use and improve outcomes.
Modeling the Physical Twin
In todayâs technological landscape, the integration of Artificial Intelligence (AI) with digital twins is gaining widespread attention and driving diverse applications. AI technologies have the potential to enhance the intelligence and autonomy of digital twins. For example, compared to traditional methods based on physics, the physics-informed AI system can significantly improve digital twins in numerous ways, such as automating the modeling process and improving computational efficiency . Furthermore, by learning from sensor data of physical systems and simulation data, AI can provide more effective and efficient predictions and fault detection . In addition, in the past two years, generative AI and Large Language Models (LLMs) have profoundly impacted digital twins, particularly in tasks related to simulation. The ongoing advancements in AI continue to push the boundaries of what digital twins can achieve, heralding a new era of smart, interconnected systems across various industries.
Physical System Modeling
Physics knowledge has long been a foundation in traditional digital twins , providing essential tools for understanding and predicting complex systems through mathematical representations of physical laws. Physics-informed AI effectively addresses a range of limitations inherent in traditional physics-based methods by combining AI techniques with physical knowledge . In this section, we review physics-based methods used in digital twins, highlight their limitations, and summarize new insights from recent physics-informed AI models.
style="width:90.0%" />
Fundamental Physics-Based Methods
Before the advent of AI, researchers predominantly relied on physics-based methods for tasks such as simulation, prediction, analysis, and control in digital twins, with computational outcomes derived from numerical methods. In this section, we will introduce several important fields in these paradigms, beginning with constructing Partial Differential Equations (PDEs) , followed by solving these PDEs through numerical methods , and finally evaluating the error through uncertainty quantification .
Partial Differential Equations (PDE).  Computational physics has long been a foundational discipline in building traditional simulations for digital twins, focusing primarily on numerical solutions for PDEs. For instance, in constructing simulators for fluid phenomena, it is often necessary to solve the Navier-Stokes equations . These equations have been extensively utilized in simulators designed for aerospace , aircraft , weather , and oil pipelines , all of which involve fluid dynamics. Similarly, in solid physics and materials science, heat equations are used to simulate the heat conduction process, and plasticity equations are used to simulate stress variations within materials. These simulations are then applied in the creation of virtual twins for architecture , manufacturing , and other industries . In addition, reaction-diffusion equations are commonly used to model diffusion processes of substances or signals, such as biochemical reactions and tumor growth . These equations play a crucial role in the simulation of digital twins, especially within the health industry and for biological entities . In addition to describing natural phenomena, PDEs are also commonly used to model various social phenomena. For instance, the spread of infectious diseases is often described by the Susceptible-Infectious-Recovered model ; traffic flow models are employed to represent traffic conditions and manage congestion prediction . In economics and finance, investment strategies and option pricing are frequently described using the Hamilton-Jacobi-Bellman equation and the Black-Scholes equation, respectively.
Numerical Methods.  To accurately obtain numerical solutions for equations in computational physics, it is necessary to perform a discretization process. This is done by dividing the domain into a discrete grid and subsequently approximating the solutions at these grid points . Based on this strategy, several numerical methods have been developed, each tailored to meet specific simulation challenges. The finite difference method simplifies implementation by using the difference of function values to approximate derivatives at grid points; however, it often lacks precision and efficiency . Spectral methods , on the contrary, represent the solution globally using different basis functions, offering high resolution (i.e., spectral accuracy ). Despite their high precision, spectral methods are generally suitable only for relatively regular domains, limiting their applicability to other downstream tasks . The finite element method (FEM) segments the domain into various small elements, such as triangles or tetrahedrons, and approximates the solution within each using low-order polynomials. Due to its flexible meshing strategy, FEM can adapt to a wide range of scenarios and complex real-world conditions . A notable example of FEM applied in simulation is the ICOsahedral Nonhydrostatic (ICON) model , which initially discretizes the Earth into an icosahedron consisting of 20 triangular faces, facilitating the calculation of numerical solutions. The ICON model demonstrates the power of FEM by converting a set of partial differential equations, which describe various weather conditions, into algebraic equations and solving them using a supercomputer. This approach has been integrated into the Earth2 system by NVIDIA.
Uncertainty Quantification.  Simulation in digital twins aims to achieve a precise one-to-one correspondence between a physical system and its virtual representation , demonstrating the importance of quantifying the uncertainties inherent in both physical measurements and computational models . The primary techniques used for quantifying uncertainty and its propagation in simulators include Monte Carlo methods , Bayesian Inference , and Sensitivity Analysis . Monte Carlo methods involve random sampling of input parameters to create a distribution of possible outcomes, enabling estimation of output uncertainties. It is particularly effective in handling complex and high-dimensional problems, which are further enhanced with variance reduction techniques such as importance sampling and stratified sampling , focusing on the most critical parts of the input space . Bayesian Inference, on the other hand, uses Bayesâ Theorem to update the probability distribution of model parameters based on prior knowledge and new data. This approach provides a systematic framework for incorporating uncertainty into model predictions, making it possible to refine the model as more data becomes available. Techniques like Markov Chain Monte Carlo are crucial in this context, as they allow the practical application of Bayesian inference by approximating posterior distributions in high-dimensional spaces . Moreover, Sensitivity Analysis evaluates how variations in input parameters affect model outputs, and identifies the key parameters that significantly influence the results. Global sensitivity analysis methods , such as the Sobol index , offer a comprehensive understanding of how input uncertainties propagate through the model, thus highlighting the most influential parameters and guiding efforts to reduce uncertainty .
Physics-Informed AI Models
In recent years, the rapid advancement of AI technologies has introduced innovative methodologies to address modeling and computation challenges in traditional physics-based methods. One approach involves using AI to explicitly extract underlying PDEs from data for further modeling process . Alternatively, another strategy involves embedding some or all of the known physical knowledge directly into AI models . Additionally, AI-assisted computation can reduce computational costs by enhancing the numerical computation steps with AI technologies . This integration streamlines processes and boosts efficiency in modeling and computation tasks.
PDEs Discovery for Modeling.  In the traditional digital twins modeling process, a significant challenge arises when certain physical laws are unclear or only partially understood, making it difficult to establish accurate mathematical models . This situation is particularly common in complex systems, such as turbulence , multiphase flows , and materials science . To address these modeling challenges, researchers have begun to explore the use of AI techniques to directly learn underlying physical knowledge and models from data. One of the earlier and highly influential works is by Brunton et al. , who introduced the Sparse Identification of Nonlinear Dynamics (SINDy) method, which can discover governing equations from time-series data. This algorithm is based on the assumption that physical laws are often simple, leading to sparsity in the indices. Using the concept of the Koopman Operator , SINDy transforms the problem of finding low-dimensional governing equations into a high-dimensional linear regression problem. A series of subsequent studies have expanded the applicability of SINDy to a wider range of scenarios, including the discovery of PDEs , handling noisy data , handling multiscale physics , and jointly processing control inputs . In particular, recent research has begun to utilize SINDy to construct simulators for digital twins in various application domains, such as manufacturing , chemical engineering , and other industrial scenarios. More recently, researchers have continued to advance the field of PDE discovery from data by utilizing deep learning. Early influential works include the numerical method-based PDE-Net and the symbol regression-based approaches followed two parallel trajectories. Further studies, such as PDE-Net 2.0 , have mixed numerical methods with symbolic approaches. Additionally, a series of works have combined deep learning techniques to further enhance the applicability and dimensionality of the SINDy algorithm.
Solving PDEs for Simulation.  Besides utilizing AI to explicitly uncover the underlying physical knowledge and mathematical models from data, another approach is to embed some or all of the known physical knowledge based on PDEs into AI models, aiming to combine the flexibility of data-driven methods with the interpretability of physical models to solve PDEs. Following this guidance, the most well-known work is the Physics-Informed Neural Networks (PINNs) . By directly incorporating PDEs into the loss function as penalty terms, PINNs allow physical information to constrain the neural network outputs to some extent.
\begin{equation}
\mathcal{L}_{\text{PINN}} = \lambda_{\text{data}} \mathcal{L}_{\text{data}} + \lambda_{\text{PDE}} \mathcal{L}_{\text{PDE}}
\end{equation}
where $`\lambda_{\text{data}}`$, $`\lambda_{\text{physics}}`$ are weights that balance the contributions of each term. The loss term for PDEs often consists of two independent parts: one part satisfies the physical equations, and the other part satisfies the boundary conditions.
\begin{equation}
\mathcal{L}_{\text{PDE}}=\mathcal{L}_{\text{physics}}+\mathcal{L}_{\text{boundary}}=\frac{1}{M} \sum_{j=1}^{M} \left( \mathcal{N}[u](x_j, t_j) \right)^2+\frac{1}{P} \sum_{k=1}^{P} \left( \mathcal{B}[u](x_k, t_k) - g(x_k, t_k) \right)^2
\end{equation}
where $`\mathcal{N}[u](x_j, t_j)`$ is the residual of the PDE at collocation points $`(x_j, t_j)`$. And $`\mathcal{B}[u] = g`$ is the boundary condition. Subsequently, researchers developed PINN variants, each enhancing PINNsâ performance and applicability from different perspectives. Variational PINNs introduced a variational formulation to improve training stability and accuracy, Conservative PINNs ensured the conservation of physical quantities, and Adaptive PINNs employed adaptive activation functions to enhance learning efficiency. Probabilistic PINNs incorporated probabilistic models to quantify uncertainty in the predictions. PINNs and their variants have been applied in a wide range of simulation domains, such as fluid dynamics , structural mechanics , and biomedical engineering .
Currently, Neural Operators have emerged as another class of methods for embedding physical knowledge. The earliest work is DeepONet , which leveraged the universal approximation theorem to directly learn differential operators in PDEs, rather than the solutions to PDEs. This approach sparked the concept of Neural Operators. Further advancements led to the Fourier Neural Operator (FNO), which significantly improved performance by filtering out high-frequency information through multiple Fourier layers. An even more advanced development based on the FNO structure is the FourCastNet , which is used for weather forecasting and has been integrated into the Earth2 system. Other variants based on FNO include Adaptive FNO , Multiwavelet Fourier Feature Operators (MWFF) . Both Neural Operators and PINNs focus the majority of computational effort during the model training phase. Once trained, these physics-informed AI models can generate predictions at remarkable speeds, often orders of magnitude faster than conventional numerical solvers . This shift in computational paradigm offers significant advantages in scenarios requiring repeated simulations or real-time predictions. For instance, Lu et al. demonstrated that their physics-informed DeepONet could solve partial differential equations up to 1000 times faster than traditional numerical methods. Similarly, Hennigh et al. showed that AI-based turbulence models could accelerate CFD simulations by up to two orders of magnitude. This dramatic reduction in inference time not only enables rapid what-if analyses and design optimizations but also opens up new possibilities for real-time control and decision-making in complex systems . However, it is important to note that the training process for these AI models can be computationally intensive and may require significant amounts of data or carefully designed loss functions incorporating physical constraints. However, the potential for rapid and accurate predictions makes physics-informed AI models an increasingly attractive option for a wide range of tasks for digital twins.
Observational Data Integration
A key challenge in modeling the physical twin is ensuring that real-world observations can be effectively incorporated into the model. This process unfolds in two steps. Section 3.2.1 addresses acquisition and alignment, where heterogeneous data from sensors, IoT devices, or logs are cleaned, synchronized, and transformed into consistent observational evidence. Section 3.2.2 then focuses on data assimilation, the stage where these prepared observations are combined with the model to update its states and parameters. In other words, acquisition and alignment ensure that the data are trustworthy and comparable, while assimilation ensures that the model itself adapts to the incoming evidence. Together, they form the methodological bridge that keeps the physical twin closely tied to the evolving reality.
Acquisition and Alignment
The construction of digital twins begins with the acquisition of observational data, which forms the bridge between physical systems and computational models. In practice, data originate from diverse sources, and their heterogeneity introduces challenges of noise, missing values, inconsistent sampling, and semantic mismatches. Acquisition and alignment methods aim to transform these disparate inputs into clean, reliable, and interoperable sequences that can serve as evidence for downstream modeling and assimilation.
Sensor Data.  The most direct form of observational data is produced by physical sensors that measure variables such as temperature, pressure, vibration, voltage, location, audio, or video. Raw sensor signals are rarely usable without processing, as they are often affected by noise, baseline drift, or intermittent dropout. Signal processing techniques such as low-pass filtering and detrending are standard approaches for mitigating these issues . Anomaly detection is also necessary to identify faulty measurements or unexpected spikes, which may otherwise corrupt the dataset. A comprehensive survey of statistical and machine learning methods for anomaly detection is provided by Hodge and Austin . Missing values caused by sensor dropout are typically reconstructed through interpolation or imputation methods, ranging from classical statistical approaches to probabilistic models . Together, these preprocessing steps convert raw sensor feeds into stable sequences suitable for integration.
IoT and Edge Data.  Beyond dedicated sensors, observational data are increasingly collected through the Internet of Things (IoT), which refers to distributed devices connected via wireless communication networks such as Wi-Fi, Bluetooth, Zigbee and LTE/5G. IoT infrastructures generate vast quantities of heterogeneous data, ranging from environmental readings to user interactions . Their communication is usually based on lightweight protocols such as MQTT or CoAP, designed for constrained environments. However, large-scale IoT deployments raise challenges including latency, packet loss, and inconsistent device configurations. To mitigate these problems, edge computing has emerged as a complementary paradigm: instead of sending all raw data to the cloud, computation is partially performed on gateways or embedded processors close to the data source . This strategy reduces bandwidth usage and response time while allowing local preprocessing such as compression or anomaly detection. More recently, the integration of AI models into edge devices has enabled real-time feature extraction and adaptive decision support, a direction often referred to as edge intelligence .
Log and Event Data.  In addition to continuous signals, digital twins incorporate discrete logs and event records that capture system-level behaviors. Logs may include textual records from control systems, operational software, or user interactions. Event data typically represent discrete occurrences, such as failures or maintenance actions. Unlike sensor and IoT data, logs are semi-structured or unstructured and require dedicated parsing methods. Log parsing frameworks such as Drain convert free-text records into structured templates suitable for analysis . A recurring challenge is inconsistent timestamps across distributed systems, which necessitates synchronization and normalization . Once standardized, logs and events can be serialized and aligned with continuous data streams, enabling their integration into digital twin workflows. These data forms highlight the importance of sequence modeling, anomaly detection, and temporal reasoning in AI-based integration approaches.
Alignment of Heterogeneous Data.  Regardless of source, acquired data streams must be aligned to ensure interoperability. Temporal alignment is the first requirement: sensors may operate at different sampling rates, IoT devices may report intermittently, and logs may record discrete events. Synchronization protocols such as the Network Time Protocol (NTP) establish a common reference clock for distributed devices . When perfect synchronization is not possible, interpolation and resampling methods are applied to harmonize signals on a shared timeline. Sliding-window aggregation can further reconcile high-frequency sensor data with lower-frequency records. Beyond time, semantic and unit normalization is necessary to prevent conflicts across heterogeneous datasets. Ontology-based mapping and dictionary-driven label alignment are commonly used for harmonizing variable names, while unit conversion ensures comparability of physical quantities such as temperature or flow rate . Together, these procedures transform heterogeneous sources into coherent datasets ready for integration with digital twin models.
Acquisition and alignment provide the methodological foundation for digital twin data integration. Sensor data, IoT and edge platforms, and log or event systems all contribute valuable but heterogeneous evidence. Through preprocessing, synchronization, and semantic normalization, raw observations are transformed into harmonized sequences that preserve fidelity while achieving interoperability. These steps also highlight opportunities for AI researchers, such as anomaly detection for noisy signals, lightweight models for edge-level feature extraction, and semantic alignment driven by representation learning.
Data Assimilation
Once observational data have been collected and aligned, the next step is to integrate them with the model in order to estimate the true system state and calibrate unknown parameters. This process, known as data assimilation (DA), provides a principled framework for merging model predictions with observational evidence. In the context of digital twins, DA is critical: without it, the model would drift away from reality due to imperfect initial conditions, incomplete parameterization, or unmodeled disturbances; with DA, the model remains dynamically consistent with the physical twin. Put simply, acquisition and alignment make the data usable, while assimilation makes the model responsive to the data. Formally, DA can be seen as a Bayesian estimation problem, where the model forecast provides a prior, the observations provide a likelihood, and the assimilation step produces a posterior estimate of the system state. Beyond updating the state, DA also delivers estimates of uncertainty, which are essential for downstream tasks such as prediction, optimization, and decision-making. Over the past decades, DA methods have evolved along three major lines: sequential approaches, variational approaches, and more recently, hybrid and learning-based approaches.
Sequential Data Assimilation.  Sequential methods update the system state step by step as new data arrive. The classical example is the Kalman Filter (KF), which provides an optimal linear estimator under Gaussian assumptions . To handle nonlinear dynamics, the Extended Kalman Filter (EKF) approximates the system by local linearization, while the Unscented Kalman Filter (UKF) employs deterministic sampling to capture nonlinear effects more robustly . For large-scale, high-dimensional systems such as geophysical models, the Ensemble Kalman Filter (EnKF) has become the method of choice . By representing error statistics with an ensemble of model trajectories, EnKF provides a scalable solution that is widely used in weather forecasting and ocean modeling. Sequential methods are attractive for digital twins because of their ability to process data streams in real time, although they may suffer from sampling errors or loss of variance in very high-dimensional settings.
Variational Data Assimilation.  Variational approaches cast DA as an optimization problem. The objective is to minimize a cost function that balances model fidelity with observational fit over a given time window. In 3D-Var, the optimization is performed at a single analysis time, combining the background forecast with new observations . In 4D-Var, the assimilation extends over a temporal window, allowing the system dynamics to constrain the analysis . These methods exploit all available observations within the assimilation window and yield dynamically consistent state trajectories. They are particularly effective for large-scale applications such as numerical weather prediction. However, their reliance on adjoint models and high-dimensional optimization makes them computationally demanding, which may limit their direct applicability in some digital twin scenarios.
Hybrid and Learning-Based Approaches.  Recent developments have sought to combine the strengths of sequential and variational schemes with machine learning. One active direction is to learn or approximate components of the assimilation pipeline. For example, neural networks can serve as surrogate observation operators, mapping model state variables to observation space when the true operator is highly nonlinear or computationally expensive . Another line of work integrates machine learning to improve error covariance modeling, a long-standing challenge in EnKF and variational schemes . The concept of Neural Data Assimilation has also emerged, where deep learning architectures are trained to emulate the update step directly, providing a data-driven approximation of Bayesian inference. Furthermore, differentiable programming frameworks have enabled the formulation of differentiable DA, where the assimilation process itself is embedded into a computation graph, facilitating end-to-end learning of model and assimilation parameters. PINNs, although already introduced in Section 3.1.2, can also be incorporated into assimilation frameworks as constraints that guide state estimation . These hybrid methods highlight a promising synergy: traditional DA offers a principled statistical framework, while AI methods bring flexibility, scalability, and the ability to exploit large datasets.
Data assimilation represents the methodological bridge between data and models in digital twins. Sequential methods emphasize real-time, step-by-step updating, variational methods leverage optimization over temporal windows, and hybrid methods incorporate machine learning to overcome long-standing limitations. Together, they ensure that the physical twin is not merely simulated but continuously synchronized with reality. In doing so, data assimilation transforms the digital twin into a dynamic mirror of its physical counterpart, capable of supporting reliable prediction, optimization, and decision-making in complex environments.
Mirroring the Physical Twin into the Digital Twin
Learning a distribution based on real data observations is a fundamental challenge in generative AI. Digital twins, which vary in their representation dimensions, require different frameworks for modeling. This section will go through some fundamental approaches and then cover recent state-of-the-art methods built upon for learning digital twins in 2D and 3D dimensions with dynamic modeling.
style="width:95.0%" />
Variational Autoencoders (VAEs).  VAEs are designed to learn an encoder-decoder pair that maps input data into a continuous latent space with two Gaussian proxy distributions. In such a framework, we train the encoder model to match the prior distribution of the latent variable $`\mathbf{z}`$ and train the decoder model to reconstruct the original image. In the encoder part, we employ a Gaussian proxy distribution $`q_\phi(\mathbf{z}\mid\mathbf{x})`$ parameterized by a neural network $`\phi`$ to approximate the intractable posterior distribution $`q(\mathbf{z}\mid\mathbf{x})`$. We seek to predict the mean and variance of the latent variable $`\mathbf{z}`$ from the input image $`\mathbf{x}`$, i.e., $`\mu_\phi(\mathbf{x})`$ and $`\sigma_\phi(\mathbf{x})`$, to align the encoded distribution closer to pre-defined prior distribution ($`p(\mathbf{z})=\mathcal{N}(\mathbf{z}; \mathbf{0}, \mathbf{I})`$). In the decoder part, assuming the images follow a Gaussian distribution, we essentially train the decoder $`p_\theta(\mathbf{x}\mid\mathbf{z})`$ to reconstruct the original image from the latent variable $`\mathbf{z}`$. To achieve these two objectives, we can maximize the evidence lower bound (ELBO) , which is defined as
\begin{align}
\mathcal{L}(\boldsymbol{\phi}, \boldsymbol{\theta}; \mathbf{x})
&= \mathbb{E}_{q_{\boldsymbol{\phi}}(\mathbf{z} \mid \mathbf{x})}\left[ \log p_{\boldsymbol{\theta}}(\mathbf{x},\mathbf{z}) - \log q_{\boldsymbol{\phi}}(\mathbf{z}\mid\mathbf{x}) \right], \\
&= \mathbb{E}_{q_{\boldsymbol{\phi}}(\mathbf{z} \mid \mathbf{x})}\left[\log p_{\boldsymbol{\theta}}(\mathbf{x} \mid \mathbf{z})\right] - \mathbb{D}_{\mathrm{KL}}\left(q_{\boldsymbol{\phi}}(\mathbf{z} \mid \mathbf{x}) \| p(\mathbf{z})\right),
\label{eq:elbo}
\end{align}
with the constraint $`\mathcal{L}(\phi, \theta; \mathbf{x}) \leq \log p_\theta(\mathbf{x})`$. Since the loss function in Eq. [eq:elbo] is differentiable, we can train the encoder $`\phi`$ and the decoder $`\theta`$ using gradient descent in an end-to-end manner. During inference time, we can sample latent variables from $`p(\mathbf{z})`$ and then feed them into the decoder to generate new images.
Normalizing Flows.  Normalizing flows are a powerful class of generative models that enable flexible and tractable density estimation by transforming a simple probability distribution into a more complex one through a series of invertible transformations. The core idea of normalizing flows is to start with a simple distribution, typically a multivariate Gaussian , and apply a sequence of bijective (invertible and differentiable) mappings to transform this simple distribution into one that matches the target data distribution. Specifically, let $`\mathbf{z}_0 \sim p_{\mathbf{z}_0}(\mathbf{z}_0)`$ represent a random variable drawn from a simple base distribution $`p_{\mathbf{z}_0}`$, such as a standard normal distribution. A normalizing flow applies a sequence of invertible transformations $`f_i: \mathbb{R}^d \rightarrow \mathbb{R}^d`$ for $`i = 1, 2, \dots, K`$, resulting in the transformed variable $`\mathbf{z}_K = f_K \circ f_{K-1} \circ \dots \circ f_1 (\mathbf{z}_0)`$. The probability density function of $`\mathbf{z}_K`$ under the transformation is given by the change of variables formula:
\begin{equation}
\label{eq:normalizing_flow}
p_{\mathbf{z}_K}(\mathbf{z}_K) = p_{\mathbf{z}_0}(\mathbf{z}_0) \left| \det \left( \frac{\partial f^{-1}}{\partial \mathbf{z}_K} \right) \right| = p_{\mathbf{z}_0}(\mathbf{z}_0) \prod_{i=1}^{K} \left| \det \left( \frac{\partial f_i}{\partial \mathbf{z}_{i-1}} \right) \right|^{-1},
\end{equation}
where $`\det \left( \frac{\partial f_i}{\partial \mathbf{z}_{i-1}} \right)`$ is the determinant of the Jacobian matrix of the $`i`$-th transformation. This formulation allows for exact likelihood computation, making normalizing flows highly effective for density estimation.
While normalizing flow offers several advantages, such as exact likelihood estimation and invertibility, they also have some limitations. One of the primary challenges is the trade-off between the expressiveness of the transformations and the computational complexity of calculating the Jacobian determinants . Simple transformations lead to efficient computation but may lack the flexibility to model complex distributions, whereas more complex transformations can capture intricate structures in the data but at the cost of increased computational burden. To tackle these challenges, recent researches attempt to enhance the expressiveness of normalizing flows, such as Neural Spline Flows and Residual Flows , which balance expressiveness and computational efficiency.
Generative Adversarial Neural Networks (GANs).  GANs is a kind of powerful framework (as shown in Figure 3) for learning data distributions, consisting of two main components: a generator ($`G`$) and a discriminator ($`D`$). These components are typically implemented as differentiable neural networks that map input data from one space to another. The optimization of GANs can be formulated as a min-max game between the generator and discriminator, with the following objective:
\begin{equation}
\min_G \max_D \mathbb{E}_{\mathbf{x} \sim p{\text{data}}(\mathbf {x})}[\log D(\mathbf{x})] + \mathbb{E}_{\mathbf{z} \sim p{\mathbf{z}}(\mathbf{z})}[\log (1 - D(G(\mathbf{z})))].
\end{equation}
The generatorâs goal is to create new examples that closely resemble the real data distribution, while the discriminator aims to accurately distinguish between real and generated examples. The training process reaches an equilibrium point, known as the Nash equilibrium , where the generator has effectively captured the true data distribution. However, GANs often face training instability issues due to the non-overlapping support between the real and generated data distributions. One approach to mitigate this problem is to introduce noise into the discriminatorâs input, thereby widening the support of both distributions. Wang et al. (2022) propose an adaptive noise injection scheme based on a diffusion model to stabilize GAN training. Due to the one-shot generation nature, GAN can be an efficient alternative compared to the more capable diffusion models, which require iterative multi-step and time-consuming denoising during inference.
Denoising Diffusion Probabilistic Models (DDPMs).  DDPMs use two Markov chains: a forward chain adding noise and a reverse chain removing it. The forward chain transforms data into a simple prior distribution, while the reverse chain, using neural networks, inverts this process. Given a data distribution $`\mathbf{x}_0 \sim q(\mathbf{x}_0)`$, the forward process generates a sequence $`\mathbf{x}_1, \mathbf{x}_2, \dots, \mathbf{x}_T`$ with transition kernel $`q(\mathbf{x}_t\mid\mathbf{x}_{t-1})`$. The joint distribution is the product of these transitions. The transition kernel is typically a Gaussian perturbation: $`q(\mathbf{x}_t\mid\mathbf{x}_{t-1}) = \mathcal{N}(\mathbf{x}_t; \sqrt{1-\beta_t} \mathbf{x}_{t-1}, \beta_t \mathbf{I}),`$ where $`\beta_t \in (0,1)`$ is a hyperparameter. This allows for analytical marginalization: $`q(\mathbf{x}_t\mid\mathbf{x}_0) = \mathcal{N}(\mathbf{x}_t; \sqrt{\bar{\alpha}_t} \mathbf{x}_0, (1-\bar{\alpha}_t) \mathbf{I}),`$ with $`\alpha_t = 1 - \beta_t`$ and $`\bar{\alpha}_t = \prod_{s=0}^{t} \alpha_s`$. The reverse process uses a learnable transition kernel:
\begin{equation}
p_\theta(\mathbf{x}_{t-1}\mid\mathbf{x}_t) = \mathcal{N}(\mathbf{x}_{t-1}; \mu_{\theta}(\mathbf{x}_t, t), \Sigma_{\theta}(\mathbf{x}_t, t)),
\end{equation}
where $`\theta`$ denotes model parameters. Itâs trained to match the time reversal of the forward process by minimizing KL divergence. The simplified loss function is:
\begin{equation}
\ell^{\text {simple }}_t(\theta)=\mathbb{E}_{\mathbf{x}_0, t, \epsilon_t}\left\|\epsilon_\theta\left(\mathbf{x}_t, t\right)-\epsilon\right\|_2^2.
\end{equation}
The denoising process proceeds step-by-step using a formula involving the learned noise prediction $`\epsilon_\theta`$ and Gaussian noise $`\mathbf{z}`$.
Sampling efficiency was enhanced through techniques like DDIM , which enables faster sampling by constructing non-Markovian diffusion processes. The EDM framework further improved efficiency and quality by refining the design space. Score-based models provided a unified view of diffusion models through stochastic differential equations. For discrete data, D3PMs extended diffusion models to discrete state spaces. Practical applications were advanced by works like GLIDE and Stable Diffusion , which enabled high-quality text-to-image generation. Classifier guidance and cross-attention control further improved conditional generation and editing.
Radiance Field Modeling with Implicit and Explicit Methods.  A radiance field provides a three-dimensional light distribution model that describes the interaction of light with surfaces and materials within an environment . It can be mathematically expressed as a function $`L \!:\! \mathbb{R}^5 \!\mapsto\! \mathbb{R}^+`$, where $`L(x, y, z, \theta, \phi)`$ denotes the mapping of a spatial point $`(x, y, z)`$ and a direction given by spherical coordinates $`(\theta, \phi)`$ to a non-negative radiance value. Radiance fields are typically represented in two forms: implicit or explicit, each offering distinct advantages for scene depiction and rendering . An implicit radiance field models the light distribution within a scene indirectly without defining the sceneâs geometry explicitly . In the context of deep learning, this often involves employing neural networks to learn a continuous representation of the volumetric scene . A notable example is NeRF , where a neural network, often a multi-layer perceptron (MLP), maps spatial coordinates $`(x,y,z)`$ and viewing directions $`(\theta, \phi)`$ to corresponding color and density values . The radiance at any point is dynamically computed by querying the MLP rather than being stored directly. This approach provides a compact and differentiable representation of complex scenes but typically requires significant computational resources during rendering due to the need for volumetric ray marching . Conversely, an explicit radiance field explicitly encapsulates the light distribution using discrete spatial structures like voxel grids or point sets . Each component of this structure encodes the radiance data for its specific spatial location, facilitating faster and more direct radiance retrieval, albeit at the expense of increased memory demands and reduced resolution .
3D Gaussian Splatting (3DGS).  3DGS  integrates the benefits of both implicit and explicit radiance fields through the use of adjustable 3D Gaussians. This method, optimized by multi-view image supervision, provides an efficient and flexible representation capable of accurately depicting scenes. It merges neural network-based optimization with structured data storage, targeting real-time, high-quality rendering with efficient training, especially for intricate and high-resolution scenes. The 3DGS model is described as:
\begin{equation}
L_{\text{3DGS}}(x, y, z, \theta, \phi) = \sum_{i} G(x, y, z, \bm{\mu}_i, \bm{\Sigma}_i) \cdot c_i(\theta, \phi),
\end{equation}
where $`G`$ symbolizes the Gaussian function, defined by the mean $`\bm{\mu}_i`$ and covariance $`\bm{\Sigma}_i`$, and $`c`$ indicates the view-dependent color. Specifically, learning 3D GS involves two main processes as depicted in Figure 3:
-
Rendering: The rendering process in 3D GS differs significantly from the volumetric ray marching used in implicit methods like NeRF. Instead, it employs a splatting technique that projects 3D Gaussians onto a 2D image plane. This process involves several key steps. First, in the frustum culling step, Gaussians outside the cameraâs view are excluded from rendering. Then, in the splatting step, 3D Gaussians are projected into 2D space using a transformation involving the viewing transformation and the Jacobian of the projective transformation. In the final step, the color of each pixel is computed using alpha compositing, which blends the colors of overlapping Gaussians based on their opacities.
-
Optimization: To achieve real-time rendering, 3D GS employs several optimization techniques, including the use of tiles (patches) for parallel processing and efficient sorting of Gaussians based on depth and tile ID. The learning process in 3D GS involves optimizing the properties of each Gaussian (position, opacity, covariance, and color) as well as controlling the density of Gaussians in the scene. The optimization is guided by a loss function combining L1 and D-SSIM losses:
MATH\begin{equation} \mathcal{L} = (1-\lambda)\mathcal{L}_1 + \lambda\mathcal{L}_{\text{D-SSIM}}, \end{equation}Click to expand and view morewhere $`\lambda`$ is a weighting factor. The density of Gaussians is controlled through point densification and pruning processes. Densification involves cloning or splitting Gaussians based on positional gradients, while pruning removes unnecessary or ineffective Gaussians.
Since the introduction of 3D Gaussian Splatting for real-time radiance field rendering , several significant advancements have been made in the field. Plenoxels introduced a neural network-free approach for photorealistic view synthesis using a sparse 3D grid with spherical harmonics. Dynamic 3D Gaussians extended the concept to dynamic scenes, enabling six-degree-of-freedom tracking and novel-view synthesis. Animatable and Relightable Gaussians focused on high-fidelity human avatar modeling from RGB videos. Other notable works include Floaters No More , which addressed background collapse in NeRF acquisition, and 3DGS-Avatar , which achieved real-time rendering of animatable human avatars. These advancements highlight the ongoing evolution and diversification of techniques in the realm of 3D scene representation and rendering.
Simulator Building
Foundational Modeling and State Representation
A digital twin simulator must first decide how to represent the physical system it seeks to emulate. In some cases, the simulator requires an explicit geometric description to reproduce spatially governed behaviors such as deformation, fluid flow, or molecular interactions. In other cases, geometry is unnecessary, and the system can be modeled through relational or temporal dependencies that capture how components interact or evolve over time. These two complementary perspectives lead to two types of state representation: geometry-based, which encodes the physical structure of the system, and abstract, which focuses on the data-driven or relational aspects of system behavior.
Geometry-Based State Representation.  In geometry-based representations, the physical system is modeled through explicit spatial constructs that define its shape, structure, and material properties. These representations form the backbone of digital twins in domains where spatial configuration directly governs system dynamics and behavior, such as engineering design, manufacturing, and biomedical modeling. They answer fundamental questions about the physical world: What does the object look like? How are its components connected? What are its physical properties?
Shape and Geometry. âWhat is the shape of an object?â - This question defines the geometric configuration of the physical entity in space. Shape representation establishes the foundation for all subsequent modeling, determining how the object occupies and interacts with its environment. Common formulations include meshes used in finite element and finite volume methods (FEM/FVM), which discretize continuous domains for numerical analysis ; point clouds capturing dense surface samples for 3D reconstruction and inspection ; voxel or grid representations used in volumetric modeling and medical imaging ; and parametric models such as CAD and Building Information Models (BIM) that encode both geometry and semantics . Neural fields, including Neural Radiance Fields (NeRF) , Signed Distance Fields (SDF) , and occupancy networks , further generalize geometry into continuous implicit functions, offering differentiable and compact scene representations. Together, these formulations provide the means to reconstruct, visualize, and simulate the spatial state of the physical world.
Topology and Connectivity. âHow are its components connected?â Beyond local geometry, topology describes the relationships and connectivity that define structural integrity and motion constraints. In mechanical systems, mesh connectivity specifies adjacency relations between elements and supports stress or deformation analysis . In articulated systems such as robotic manipulators, kinematic chains describe hierarchical dependencies and degrees of freedom among joints and links . Assembly graphs further encode how individual parts interface or move relative to one another, enabling simulation of multi-body dynamics and structural coupling . Accurate topological modeling ensures that the digital twin preserves the structural logic of the physical system, allowing analyses such as load propagation, collision detection, and deformation tracking under physical constraints.
Physical Attributes. âWhat are its physical properties?â Geometry and topology alone cannot determine the systemâs behavior without describing its intrinsic material and boundary properties. These attributes define how the physical entity responds to external forces, heat, or other environmental stimuli. Key descriptors include material properties such as density, elasticity, and viscosity; boundary conditions specifying loads, fixed supports, or fluid interfaces; thermal parameters including conductivity and specific heat; and initial conditions that define the starting state of fields such as temperature or velocity. Together, these quantities enable accurate numerical simulation and predictive modeling through governing physical equations. The inclusion of such physical attributes transforms geometric models from static visualizations into dynamic computational entities that mirror the real-world systemâs mechanical, thermal, or electromagnetic responses.
Non-Geometric State Representation.  Not all simulators require explicit geometry to capture system behavior. In many digital twins, the key dynamics arise from relationships, temporal patterns, or statistical dependencies rather than spatial form. Non-Geometric State Representations therefore describe systems through symbolic, relational, or feature-based structures, enabling modeling of processes where geometry is unavailable, irrelevant, or computationally unnecessary. Such representations are widely used in cyber-physical, social, and biological systems to support large-scale reasoning and prediction.
Feature-based Representation. Feature-based representations encode complex system states into compact numerical vectors, allowing efficient computation and integration across modalities. These embeddings can be extracted from text, images, and sensor measurements, or learned through self-supervised and multimodal models . Large-scale foundation models have further generalized this concept by learning unified latent spaces capable of representing diverse system behaviors and attributes. Within digital twin simulators, embeddings provide scalable interfaces for high-dimensional inference, supporting fast querying, cross-domain adaptation, and intelligent decision-making.
Time-Series Representation. Time-Series representations describe how system states evolve over time. They are central to digital twins that monitor and forecast behaviors from continuous data streams, such as sensor signals, physiological measurements, or environmental variables . Recurrent and attention-based models, including RNNs, TCNs, and Transformers , have shown strong ability to capture both short- and long-term dependencies. In healthcare twins, these representations enable dynamic patient monitoring and disease progression modeling ; in industrial twins, they support predictive maintenance and anomaly detection based on telemetry data .
Graph-based Representation. Graph-based representations model a system as a collection of entities and their interactions, providing a natural framework for describing relational dynamics. This formulation has proven effective for networked infrastructures such as transportation systems , supply chains , and energy grids , as well as for semantic and biomedical knowledge graphs . Graph Neural Networks (GNNs) extend these ideas by learning representations over structured dependencies, enabling simulation of flow, fault propagation, and system-level optimization.
Non-Geometric State Representations thus complement geometry-based models by focusing on relationships and dynamics rather than spatial fidelity. Together, they define two foundational paradigms for simulator construction: one grounded in physical form, the other in data and interaction. Modern digital twins often integrate both perspectives. For example, structural components can be modeled geometrically, while control, communication, or biological processes are represented through graphs or embeddings. This hybrid approach enables simulators to capture both the physical dynamics and semantic interactions of complex systems.
Behavior and Process Simulation
State-Space Simulation.  State-space simulation represents one major direction for building simulators. The goal is to learn system behavior directly from data, treating simulation as function approximation that maps inputs to outputs or as sequence prediction that forecasts temporal evolution. Rather than solving governing equations explicitly, these models infer the underlying dynamics from observational evidence, allowing digital twins to emulate complex processes efficiently when analytical formulations are unavailable or computationally prohibitive .
Feedforward models such as MLPs, CNNs, and ResNets learn steady-state mappings between design or control variables and resulting performance indicators. They are often used as differentiable surrogates for engineering optimizationâfor instance, predicting aerodynamic lift from airfoil geometry, estimating heat dissipation from material parameters, or forecasting drug release rate from formulation properties . By replacing expensive finite-element simulations, such neural surrogates enable rapid design iterations and sensitivity analysis. When temporal dependencies dominate, recurrent and attention-based networks including LSTMs, GRUs, and Transformers capture the dynamic evolution of system states. These models treat simulation as sequence forecasting, learning how a process unfolds over time. In industrial digital twins, they are applied to predict sensor trajectories for fault detection ; in healthcare, they model disease progression using patient time-series data ; in transportation, they forecast multi-step traffic flow and congestion propagation across urban networks .
For systems with explicit spatial or relational structure, GNNs extend this paradigm by learning interactions among interconnected components. Each node represents a physical or logical entity, and edges describe dependencies such as force transmission or resource exchange. This allows GNN-based simulators to reproduce mesh deformation in mechanical systems, voltage propagation in power grids, and flow redistribution in transportation networks. Through message passing, GNNs enable digital twins to model how local perturbations collectively shape global dynamics. Beyond discrete representations, neural operators such as DeepONet and the Fourier Neural Operator (FNO) generalize state-space learning to functional mappings. Instead of approximating a single trajectory, they learn solution operators of partial differential equations, directly mapping boundary conditions or source terms to entire solution fields. Once trained, these operators provide fast and high-fidelity surrogates for simulating fluid flow, heat transfer, and material deformation , achieving speedups of several orders of magnitude over traditional solvers.
In summary, state-space simulation focuses on learning how systems respond and evolve, either by approximating steady mappings or by predicting temporal sequences. From predicting equipment failure to optimizing engineering designs and accelerating physical simulations, this approach forms a central computational pathway for building intelligent digital twins that learn, adapt, and generalize across physical domains.
Visual World Simulation.  Recent advances in generative AI have introduced a new paradigm for simulation by directly generating realistic visual observations of the world rather than explicitly modeling its internal states. In this view, predicting future frames or synthesizing dynamic scenes becomes equivalent to simulating the evolution of the physical environment. Such world simulators enable embodied agents, including robots and autonomous vehicles, to learn, plan, and interact within controllable, data-driven virtual worlds without relying solely on costly physical experiments.
Video diffusion models have emerged as one of the most powerful frameworks for generative simulation. Leveraging large-scale video datasets and diffusion-based architectures, these models can synthesize high-fidelity and temporally coherent videos that approximate real-world physics and dynamics. For instance, VideoComposer introduces a compositional video synthesis framework conditioned on text, spatial layout, and temporal cues, allowing precise control over generated motion and scene composition. DynamiCrafter extends this capability by animating still images through motion priors learned from text-to-video diffusion models, effectively turning static scenes into dynamic simulations. Most notably, Sora demonstrates the remarkable potential of text-to-video diffusion models as universal world simulators, capable of generating physically consistent, photorealistic videos from natural language descriptions.
While realism is critical, controllability remains essential for simulation-driven learning. Several recent studies have focused on integrating structured control into video generation models to produce task-relevant, interactive environments. Seer introduces a frame-sequential text decomposer that translates global instructions into temporally aligned sub-instructions, enhancing fine-grained control over the resulting video trajectories. Video Adapter provides a lightweight adaptation mechanism for large pre-trained video diffusion models, enabling efficient domain customization without full finetuning. At a larger scale, the Cosmos World Foundation Model Platform offers an integrated framework for constructing controllable video-based world models, including video curation pipelines, tokenizers, and pre-trained foundation simulators for physical AI applications such as robotics and autonomous driving.
Beyond pure video generation, interactive world models like Genie and its successor Genie 2 extend generative simulation into embodied, action-controllable 3D environments. These systems can generate playable, open-ended virtual worlds conditioned on text, sketches, or other multimodal prompts, enabling autonomous agents to learn and act within dynamic visual environments.
Video generation as world simulation thus represents a paradigm shift from explicit physics-based modeling to observation-driven synthesis. By learning the visual and temporal structure of reality, such models offer controllable, scalable, and photorealistic environments for training, testing, and reasoning. Within digital twin systems, they bridge simulation and perception, allowing virtual agents to both observe and interact with realistic representations of the physical world.
Simulator Visualization
Scene Modeling
In digital twins, scene modeling supports visualization by defining how the simulated world appears in three-dimensional space. Once the simulator engine has been built to model system behavior, visualization focuses on reconstructing or synthesizing the worldâs visible structure. Scene modeling determines what the twin âlooks like.â Scene modeling aims to represent the spatial configuration and visual appearance of real or virtual environments by learning how light interacts with matter through radiance field representations. With advances in neural rendering, data-driven methods can now recover highly detailed and spatially consistent 3D and even 4D scenes directly from multi-view images or sensor data. This section introduces two main directions in this field: large-scale static scene reconstruction and dynamic scene modeling.
Static Scene Reconstruction.  Scaling neural rendering techniques to large urban environments has been a significant focus of recent research. These methods aim to capture the complexity of city-scale scenes while maintaining high visual fidelity and efficient rendering. Block-NeRF introduces a variant of NeRF that can represent large-scale environments by decomposing the scene into individually trained NeRFs. This approach decouples rendering time from scene size, enabling rendering to scale to arbitrarily large environments. Urban Radiance Fields extends NeRF to handle asynchronously captured lidar data and address exposure variation between captured images, producing state-of-the-art 3D surface reconstructions and high-quality novel views for street scenes. TensoRF proposes modeling the radiance field as a 4D tensor and introduces vector-matrix decomposition to achieve fast reconstruction with better rendering quality and smaller model size compared to NeRF. NeRF in the Wild addresses the challenges of unconstrained photo collections, enabling accurate reconstructions from internet photos of famous landmarks. More recent works have focused on further improving the scalability and quality of large-scale scene reconstruction. K-Planes introduces a white-box model for radiance fields using planes to represent d-dimensional scenes, providing a seamless way to go from static to dynamic scenes. BungeeNeRF achieves level-of-detail rendering across drastically varied scales, addressing the challenges of extreme multi-scale scene rendering. Global-guided Focal Neural Radiance Field proposes a two-stage architecture to achieve high-fidelity rendering of large-scale scenes while maintaining scene-wide consistency. CityGaussian employs a novel divide-and-conquer training approach and Level-of-Detail strategy for efficient large-scale 3D Gaussian Splatting training and rendering.
Dynamic Scene Modeling.  Extending static representations to model dynamic scenes with moving objects has been another important direction in neural rendering research. These methods aim to capture both the spatial and temporal aspects of complex real-world environments. 4D Gaussian Splatting approximates the underlying spatio-temporal 4D volume of a dynamic scene by optimizing a collection of 4D primitives, enabling real-time rendering of complex dynamic scenes. Scalable Urban Dynamic Scenes (SUDS) introduces a factorized scene representation using separate hash table data structures to efficiently encode static, dynamic, and far-field radiance fields. Street Gaussians proposes a novel pipeline for modeling dynamic urban street scenes, using a combination of static and dynamic 3D Gaussians with optimizable tracked poses for moving objects. Deformable 3D Gaussians introduces a method that reconstructs scenes using 3D Gaussians and learns them in canonical space with a deformation field to model monocular dynamic scenes. DynMF presents a compact and efficient representation that decomposes a dynamic scene into a few neural trajectories, allowing for real-time view synthesis of complex dynamic scene motions. Multi-Level Neural Scene Graphs proposes a novel, decomposable radiance field approach for dynamic urban environments, using a multi-level neural scene graph representation that scales to thousands of images with hundreds of fast-moving objects.
Interactive Visualization and Interfaces
While scene modeling establishes the structural and visual substrate of the virtual environment, interactive visualization determines how humans and AI agents see, explore, and make sense of that environment. This layer transforms simulated or reconstructed worlds into perceivable, analyzable views, supporting situational awareness, hypothesis testing, and collaborative understanding through intuitive visual interfaces.
Immersive and Real-Time Visualization.  Advances in rendering pipelines and GPU acceleration now allow digital twins to achieve photorealistic, real-time visualization of large-scale environments. Neural rendering techniques such as 3D Gaussian Splatting make it possible to maintain interactive frame rates without sacrificing fidelity, while compact scene representations like Instant-NGP further improve efficiency for view synthesis. Immersive visualization systems, including AR/VR headsets and CAVE displaysâbring depth perception and spatial presence to users , enabling intuitive exploration through gestures, motion tracking, or gaze-based control. Such immersive systems enhance human perception and understanding in domains such as surgical simulation, smart city management, and industrial training .
Interactive Dashboards and Visual Analytics.  Beyond immersion, digital twins rely on interactive dashboards and visual analytics tools to organize and interpret simulation results and live sensor data . These interfaces integrate predictive models, streaming information, and diagnostic views into a unified visual layer that supports real-time monitoring and reasoning . In practice, they appear as 3D monitoring dashboards for smart factories , preoperative visualization systems for surgical planning , or operational control centers that display traffic flow and energy distribution for situational analysis . Recent cloud-based visualization frameworks further allow distributed teams to collaboratively explore, annotate, and interpret digital twin environments through synchronized visual interfaces.
Intervening in the Physical Twin via the Digital Twin
Predicting Physical Behavior
Prediction modeling is a fundamental aspect of digital twin systems, allowing for the forecasting of future conditions and behaviors based on current and historical data. In digital twin systems, which create virtual replicas of physical entities, predictive models can forecast various scenarios such as equipment malfunctions , performance degradation , and system anomalies . These predictions are essential for optimizing performance, preventing unexpected issues, and ensuring the efficient operation of physical assets . Leveraging advanced AI techniques such as machine learning and deep learning methods, digital twins can analyze extensive data to predict trends , detect anomalies , and make decisions with real-time . This section will delve into two parts, focusing on: 1) Prediction Modeling in Digital Twin Systems; 2) Types of Prediction in Digital Twin Systems.
Prediction Modeling Fundamentals
Prediction modeling is a cornerstone of digital twin systems, providing the capability to forecast future states and behaviors of physical entities based on current and historical data. This capability is critical for optimizing operations, planning maintenance, and improving the overall performance and reliability of the system. This section introduces the concept of prediction modeling, provides a high-level mathematical definition, and presents relevant examples within digital twin systems.
Definition of Prediction Modeling.  Prediction modeling in digital twin system involves forecasting future states $`\mathbf{x}_{t+T} \in \mathcal{X}`$ using observed data $`\mathbf{X}_t \subseteq \mathcal{X}`$ and a prediction function $`\mathcal{F}`$. Formally, given a set of observed states $`\mathbf{X}_t`$ up to the current time $`t`$, the goal is to estimate the future state $`\hat{\mathbf{x}}_{t+T}`$ at a future time $`t+T`$ as $`\hat{\mathbf{x}}_{t+T} = \mathcal{F}(\mathbf{X}_t, T)`$, where $`\mathbf{X}_t = \{\mathbf{x}_{t_1}, \mathbf{x}_{t_2}, \ldots, \mathbf{x}_t\}`$ represents the historical data up to the current time $`t`$, and $`T`$ is the prediction horizon, representing the time interval into the future for which the prediction is made. The state space $`\mathcal{X}`$ includes all possible states of the system, encompassing various sensor readings, operational conditions, and performance metrics. The observed data $`\mathbf{X}_t`$ are the recorded states used to train the prediction models. The function $`\mathcal{F}`$ employs machine learning, deep learning or statistical methods to predict future states based on historical and current data. To optimize the predictive function $`\mathcal{F}`$, the objective is to minimize a loss function $`\mathcal{L}`$ that captures the difference between the predicted state $`\hat{\mathbf{x}}_{t+T}`$ and the actual future state $`\mathbf{x}_{t+T}`$ as:
\mathcal{L} = \mathbb{E}[\ell(\mathbf{x}_{t+T}, \hat{\mathbf{x}}_{t+T})],
where $`\ell`$ is the task-specific loss function (e.g., mean squared error $`\ell(\mathbf{x}, \hat{\mathbf{x}}) = (\mathbf{x} - \hat{\mathbf{x}})^2`$ in regression tasks).
Predictive Tasks
In digital twin systems, prediction models play a crucial role in maintaining the reliability , efficiency , and performance of both physical and cyber components. By leveraging advanced AI techniques, these models enable continuous monitoring and analysis of complex data , identifying patterns and forecasting potential issues before they manifest. The primary types of prediction in digital twin systems can be broadly categorized into two main areas: 1) Real-time Decision Making; 2) Predictive Maintenance. The following sub-sections will delve into these categories, outlining the key methods and tasks to highlight their importance and effectiveness in digital twin systems.
Real-time Decision Making.  Real-time decision-making is crucial for digital twin systems, which create virtual replicas of physical entities to enable continuous monitoring and simulation. This capability allows for immediate analysis and response using current and historical data, supporting rapid issue identification, operational optimization, and improved system efficiency. Real-time decision-making in digital twin system spans various domains, including traffic management , industrial logistics , healthcare , fire safety , structural health monitoring , energy management . The following sections explore key methodologies and tasks in these areas, highlighting how advanced AI techniques facilitate real-time decision making. a) For traffic management, Cai et al. developed a hybrid encoder-decoder neural architecture namely Traffic Transformer for traffic forecasting. A graph convolutional network is used to model spatial dependencies and a transformer is utilized to model temporal dependency in which novel temporal positional encoding strategies are proposed to encode the continuity and periodicity of time series. Li et al. developed a Multisensor Data Correlation Graph Convolution Network (MDCGCN) to address the challenges of real-time traffic flow prediction. This model effectively captures dynamic temporal and spatial correlations in traffic patterns, significantly improving the prediction accuracy. Similarly, Liang et al. proposed the Spatial-Temporal Aware Data Recovery Network (STAR), which uses graph neural networks to recover missing entries in spatial-temporal traffic data. This approach ensures accurate data recovery critical for real-time traffic monitoring and decision making in Intelligent Transportation Systems (ITS). Kong et al. introduced the Dynamic Graph Convolutional Recurrent Imputation Network (DGCRIN), which models dynamic spatial dependencies and utilizes diverse data for imputing missing traffic data, thus enhancing real-time traffic data analysis. b) In industrial logistics, Wu et al. presented a system using the industrial Internet of Things and long short-term memory (LSTM) networks for real-time tracking of manufacturing resources. This system improves operational efficiency by accurately locating product trolleys and enabling location-based services. Wang and Ma designed the PhysiQ framework to monitor physical therapy exercises at home. This system uses a multi-task spatiotemporal Siamese Neural Network to measure exercise quality, ensuring patients perform exercises correctly in real time. c) In fire safety field, Zhang et al. proposed the Artificial-Intelligence Digital Fire (AID-Fire) system, which uses convolutional LSTM neural networks to identify and monitor fire evolution in real time, significantly aiding in firefighting and evacuation processes. d) In the medical field, Pan et al. developed the Temporal-based Swin Transformer Network (TSTNet) for real-time surgical video workflow recognition, achieving high accuracy by modeling temporal information and multi-scale visual data. e) In battery management field, Li et al. presented a framework combining convolutional neural networks and LSTM for real-time degradation prediction of lithium-ion battery degradation in real-time, ensuring accurate battery health monitoring. Similarly, Yi et al. proposed a method for real-time temperature prediction and degradation analysis of lithium-ion batteries, using LSTM networks to maintain battery safety and performance.
Predictive Maintenance.  Predictive maintenance is essential for ensuring the reliability and longevity of machinery and equipment across various industries. By utilizing advanced AI techniques, predictive maintenance enables accurate forecasting of equipment failures and optimized maintenance schedules . a) In the manufacturing sector, Li et al. introduced a hierarchical attention graph convolutional network (HAGCN) that combines spatial and temporal dependencies to predict the remaining useful life (RUL) of machinery. Similarly, Aivaliotis et al. used physics-based simulations to estimate RUL for industrial robots, demonstrating the practicality of integrating digital models with real-time data. Moreover, Luo et al. proposed a hybrid method that blends model-based and data-driven techniques for accurate CNC machine tool life prediction. b) For systems and equipment diagnostics, several methods have been introduced. Yang et al. presented SuperGraph, a spatial-temporal graph-based feature extraction method for rotating machinery fault diagnosis, which demonstrates significant advantages in handling complex data. Similarly, Zhang et al. implemented a fault prediction system for electromechanical equipment using multivariate spatial-temporal graph neural networks, which enhances predictive accuracy. In the realm of power electronics, Peng et al. developed a non-invasive health indicator estimation method for DCâDC converters, employing particle swarm optimization to monitor key components effectively. Addressing challenges within the IoT ecosystem, Altun and Tavli explored distributed ledger technologies to propose a model that improves security and scalability in predictive maintenance applications. Mubarak et al. combined machine learning and advanced analytics to create a comprehensive predictive maintenance framework for Industry 4.0, optimizing maintenance decisions and improving cost-effectiveness. c) In the aerospace industry, Liu et al. highlighted the integration of various data sources to support decision-making processes, enhancing the efficiency of predictive maintenance. Similarly, Tuegel et al. focused on reengineering aircraft structural life prediction using high-fidelity models ensuring greater accuracy and safety in assessing structural integrity.
Detecting and Diagnosing Anomaly
Anomaly detection (i.e., fault detection) is the process of identifying and responding to unusual patterns or behaviors within a system that deviates from the norm . In digital twin systems, which create virtual replicas of physical entities, anomalies can manifest in various forms such as equipment malfunctions , unexpected changes in operational performance , or irregular patterns in sensor data . These anomalies can indicate underlying issues that may lead to significant network failures , safety hazards , and financial losses if not addressed promptly. Therefore, anomaly detection is crucial for maintaining the integrity , reliability , and efficiency of the digital twin systems. By leveraging advanced anomaly detection algorithms (e.g., Autoencoders, GANs, and RNNs), digital twin systems can continuously monitor and analyze real-time sensor data and historical operational records to identify irregularities , predict potential failures , and prescribe maintenance actions , enabling timely interventions and preventive maintenance before costly breakdowns occur . This section will delve into three aspects, focusing on: 1) Anomalies in Digital Twin Systems; 2) Types of Anomaly Detection; 3) Anomaly Detection Methods.
Anomaly Characterization
Anomalies in digital twin systems are deviations from expected behavior that can indicate underlying issues in the physical or virtual components of the system . Understanding these anomalies is essential for developing effective detection and monitoring methods This section will introduce the concept of anomalies, provide mathematical definitions, and give examples relevant to digital twin systems.
Definition of Anomalies in Digital Twin Systems.  An anomaly in a digital twin system is any deviation from the expected behavior of the system, which can encompass various aspects such as data, processes, and performance metrics. Formally, given a system state space $`\mathcal{S}`$, an anomaly is defined as a state $`s \in \mathcal{S}`$ that significantly deviates from the expected behavior $`\mathcal{E} \subset \mathcal{S}`$. This can be expressed as:
s \in \mathcal{A} \iff \mathbb{P}(s \mid \mathcal{E}) < \epsilon,
where $`s`$ represents a state of the system, such as a specific temperature reading or a performance metric like the speed of a manufacturing line. The state space $`\mathcal{S}`$ includes all possible states the system can be in, encompassing every possible temperature, pressure, or performance metric. The expected behavior $`\mathcal{E}`$ represents the subset of $`\mathcal{S}`$ that includes normal operating states, such as temperature ranges between 50°C and 80°C. The set of anomalies $`\mathcal{A}`$ includes states outside this expected range, such as a temperature reading above 90°C. The probability $`\mathbb{P}(s \mid \mathcal{E})`$ measures how likely $`s`$ is to occur under normal conditions, and if this probability is lower than a predefined threshold $`\epsilon`$, the state $`s`$ is flagged as an anomaly. This broad definition encompasses various types of anomalies including data anomalies, process anomalies, and performance anomalies, each indicating different potential issues within the digital twin system.
Anomaly Detection Types and Methods
In digital twin systems, anomalies can be categorized into several types based on their characteristics and impact on the system. The following subsections describe the main types of anomaly detection required in digital twin systems: 1) Data-driven Anomaly Detection; 2) System-based Anomaly Detection.
Data-driven Anomaly Detection.  Data-driven anomaly detection is essential in digital twin systems, where massive amounts of data from various sensors and systems are continuously transmitted and monitored. Anomalies in data can arise from numerous sources, including sensor malfunctions , network issues , and operational faults . a) Sensor data anomalies are often caused by sensor failures or inaccuracies in data collection. To address this, various anomaly detection techniques have been employed. For instance, Hu et al. proposed a masked one-dimensional convolutional autoencoder (MOCAE) for bearing fault diagnosis. The model leverages deep learning methods to enhance feature extraction and improve fault detection accuracy. Similarly, Darvishi et al. introduced a deep recurrent graph convolutional architecture for sensor fault detection, isolation, and accommodation. This approach constructs virtual sensors to refurbish faulty data and uses a classifier to detect and isolate faults. Moreover, Hasan et al. developed a Wasserstein GAN-based model for early fault detection in wireless sensor networks. This model uses Gramian angular field encoding to convert time series data into images, which are then processed by a GAN to detect anomalies, achieving high fault detection accuracy. b) Network data anomalies arise from issues in data transmission and network performance (e.g., high latency), affecting the reliability and timeliness of the data received by digital twin systems. Li et al. addressed this challenge by detecting anomalies in internet service quality over fixed access networks. Their system aggregates data from multiple network elements and employs real-time simulations to detect service quality degradation and network faults. Moreover, the importance of data-centric middleware in large-scale digital twin platforms is emphasized in . Their proposed architecture supports efficient data communication within digital twin systems, which is critical for accurate anomaly detection and system reliability. c) Operational data anomalies occur due to irregularities or faults in the operational processes of systems. These anomalies can be particularly challenging to detect and diagnose due to the complex interactions between various system components. To address this challenge, many anomaly detection methods have been proposed. For example, Lian et al proposed a method for detecting anomalies in multivariate time series data from oil and gas stations. Their MTAD-GAN approach combines knowledge graph attention with temporal Hawkes attention to accurately identify and interpret operational anomalies. In the context of power systems, Shi et al. leverage random matrix theory and free probability theory for anomaly detection, accurately characterizing data correlations and effectively identifying anomalies in complex operations. Moreover, Hao et al. developed a hybrid statistical-machine learning model that integrates SARIMA and LSTM for detecting anomalies in industrial cyber-physical systems. This model effectively identifies cyberattacks, malicious behaviors, and network anomalies with high accuracy and low computational complexity, making it suitable for real-time applications.
System-based Anomaly Detection.  System-based anomaly detection in digital twin systems focuses on identifying and diagnosing faults and irregularities that arise from the overall systemâs operational processes and network communications . These anomalies can stem from various sources, such as hardware malfunctions and cyber-attacks. We categorize these anomalies into two main types: a) Operational fault diagnosis; b) Cybersecurity and network communications. a) Operational fault diagnosis involves detecting and diagnosing faults within the operational processes of systems, which can include machinery failures, process disruptions, and component wear and tear. For instance, Wu et al. developed a multi-layer convolutional neural network for real-time fault diagnosis in high-speed train bogies, enhancing safety and reducing maintenance costs. Similarly, in the manufacturing sector, Lv et al. designed a fault identification algorithm based on active learning and domain adversarial neural networks (DANN), which significantly improved fault diagnosis accuracy and stability under varying operational conditions. Ghosh et al. constructed a system using hidden Markov models to encapsulate system dynamics and enhance fault diagnosis in manufacturing processes through improved understanding, prediction, and decision-making capabilities. Additionally, Xu et al. proposed a two-phase fault diagnosis method using deep transfer learning, facilitating real-time monitoring and predictive maintenance by migrating trained models from virtual to physical spaces, thus maintaining operational continuity in dynamic production environments. Furthermore, Wang et al. explored various classification models for anomaly detection in smart manufacturing, demonstrating the effectiveness of decision trees in achieving high fault classification accuracy and preventing operational disruptions. b) Cybersecurity and network communications anomalies involve issues related to data transmission, network performance, and security threats, which can compromise the reliability and integrity of digital twin systems. To address these challenges, researchers have developed various approaches. For instance, Wang et al. utilized hidden Markov models and transfer learning to identify faulty components in virtual machines within NFV environments, enhancing fault recovery and ensuring secure network operations. Similarly, Liu et al. proposed an autonomous trusted network framework that integrates data aggregation, security models, and intelligent configuration models to proactively detect and mitigate threats, thereby preventing security breaches and data loss. Moreover, Balta et al. developed a robust defense mechanism for cyber-physical manufacturing systems. By distinguishing between expected anomalies and cyber-attacks using a combination of data-driven machine learning and physics-based models, they ensure the integrity of manufacturing processes and protect against malicious activities.
To effectively detect and respond to anomalies that appear in both cyber and physical components of digital twin systems, a variety of advanced machine learning and deep learning models have been introduced . These models enable the continuous monitoring and analysis of complex data, identifying patterns and predicting potential issues to ensure the reliability, integrity, and efficiency of the systems.
Machine Learning-based Anomaly Detection.  Machine learning-based anomaly detection methods in digital twin systems employ a range of traditional algorithms to identify irregularities across various domains. These methods encompass clustering and distance-based techniques , probabilistic approaches , as well as discriminative models , each offering unique strengths in handling high-dimensional data, capturing complex patterns, and providing robust solutions for real-time anomaly detection and fault diagnosis. a) Clustering and distance-based methods are effective for applications requiring quick and adaptive responses to anomalies. Shetve et al. propose an adaptive N-step technique that integrates DBSCAN, Isolation Forest, and Local Outlier Filter to achieve high accuracy in detecting anomalies in smart manufacturing environments. Zhang et al. enhance k-nearest neighbors (kNN) by introducing weighted distances to improve fault detection in multimodal processes. In healthcare, Sarris et al. develop a K-means-based algorithm for brain tumor detection from MRI scans, while Abirami and Karthikeyan propose an optimized fuzzy k-nearest neighbor classifier for early Parkinsonâs disease identification. b) Probabilistic-based methods provide robust solutions for managing uncertainty and complex data relationships. Ademujimi and Prabhu utilize Bayesian Networks trained through a co-simulation approach for fault diagnostics in smart manufacturing systems. Ruah et al. present a Bayesian framework for anomaly detection in wireless systems, which addresses model uncertainty to enhance detection and data optimization. In distribution networks, Shi et al. apply a random matrix theory (RMT) based method for early anomaly detection using SCADA data and a spatio-temporal correlation analysis approach for locating anomalies . Yu et al. propose a nonparametric Bayesian network for health monitoring, employing an improved Gaussian particle filter (GPF) and Dirichlet process mixture model (DPMM) for real-time updates. Maleh demonstrates the effectiveness of machine learning models in cybersecurity for IoT systems, achieving high accuracy even in constrained environments. c) Discriminative models, such as support vector machines (SVM), are effective in defining decision boundaries for classification tasks. Gaikwad et al. integrate thermal simulations and sensor data in a machine learning framework using SVM to detect process faults in additive manufacturing. Yin and Hou highlight the advantages of SVM in fault monitoring and diagnosis for complex industrial processes, emphasizing its generalization performance and suitability for small sample scenarios.
Deep Learning-based Anomaly Detection.  Deep learning-based anomaly detection techniques have significantly advanced the capabilities of digital twin systems by leveraging sophisticated neural network architectures to detect complex patterns and irregularities in various domains. These methods include autoencoders , generative adversarial networks (GANs) , convolutional neural networks (CNNs) , recurrent neural networks (RNNs) , and hybrid methods , each providing unique advantages in modeling high-dimensional data, capturing temporal and spatial dependencies, and ensuring robust anomaly detection and fault diagnosis. a) Autoencoders and their variants are particularly effective for unsupervised anomaly detection and feature extraction. Castellani et al. introduce a weakly supervised approach using Siamese Autoencoders (SAE) for industrial anomaly detection, outperforming state-of-the-art methods in various settings. Hu et al. propose a masked one-dimensional convolutional autoencoder (MOCAE) for bearing fault diagnosis, demonstrating superior performance on real bearing datasets. Similarly, Xu et al. utilize deep transfer learning with a Stacked Sparse Autoencoder (SSAE) model to achieve high-fidelity fault diagnosis in dynamic manufacturing processes. b) Generative adversarial networks (GANs) have proven effective in capturing complex data distributions and generating realistic data for anomaly detection. Xu et al. present a GAN-based approach called ATTAIN for cyber-physical systems, utilizing GCN-LSTM modules to enhance anomaly detection capabilities. Hasan et al. propose a Wasserstein GAN-based model for early drift fault detection in wireless sensor networks, achieving high accuracy in detecting sensor faults. Lian et al. introduce a digital twin-driven MTAD-GAN for multivariate time series anomaly detection in oil and gas stations, leveraging attention mechanisms to improve detection performance. c) Convolutional neural networks (CNNs) are leveraged for their ability to handle spatial data and extract hierarchical features. For example, Danilczyk et al. utilize a CNN within the ANGEL Digital Twin environment to detect and classify faults in power systems using high-fidelity measurement data. Li et al. employ a multidimensional deconvolutional network with attention mechanisms for real-time anomaly detection in industrial control systems. d) Recurrent neural networks (RNNs), particularly LSTM networks, are adept at capturing temporal dependencies in sequential data. For instance, Feng and Tian propose the NSIBF method combining neural system identification and Bayesian filtering for robust anomaly detection in cyber-physical systems. Kumar et al. integrate LSTM-SAE and BiGRU with self-attention mechanisms for secure communication in digital twin-empowered IIoT networks, enhancing intrusion detection capabilities. e) Other hybrid methods like deep recurrent graph convolutional architectures have also been utilized to address sensor fault detection and accommodation. Darvishi et al. present a deep recurrent graph convolutional architecture for sensor fault detection, isolation, and accommodation in large-scale networked systems. Additionally, hybrid approaches combining deep learning with other methods have been explored. Hao et al. develop a hybrid model integrating SARIMA and LSTM for real-time anomaly detection in ICS networks, providing high detection accuracy with low computational complexity.
Optimizing and Controlling
AI-enhanced optimization and control methods represent a significant advancement in managing complex systems, merging the capabilities of artificial intelligence with the detailed modeling and simulation afforded by digital twins . This integration results in smarter, more responsive, and predictive optimization and control mechanisms . The enhancement of AI in digital twins for optimization and control systems can be seen in the following aspects, including AI-enhanced optimization and adaptive control.
Optimization Strategies
Traditional mathematical programming-based optimization methods for digital twins often struggle to handle scenarios requiring rapid response or complex environmental changes that demand continuous adaptivity . AI-enhanced optimization methods can address these issues to some extent . This section will be divided into two parts: the first part will discuss how AI improves response speed to enhance real-time optimization for digital twins, thereby enhancing system performance and reducing resource consumption; the second part will summarize how AI enhances adaptivity, leading to better robustness and improved handling of uncertainties in adaptive optimization tasks for digital twins.
Real-time Optimization.  Real-time Optimization refers to the immediate adjustment of physical systems based on real-time sensor data and simulation results to optimize system performance. In related tasks that require immediate decision-making and fast response, AI enhances response speed and decision efficiency, improving the quality of decisions. For instance, early work presents a reinforcement learning-based real-time optimization (RL-RTO) methodology for process systems, which integrates optimal decision-making into a neural network, contrasting with traditional repeated process model solutions. The RL-RTO approach, demonstrated using a chemical reactor, shows potential by improving annual profit by 9.6%, though it still lags behind the conventional first principles plus nonlinear programming method which achieved a 17.2 % improvement. Further, Dong-Hoon Oh et al. introduce an actor-critic reinforcement learning strategy for optimizing hydrocracking unit operations, developed from a validated mathematical model with less than 2% error. The approach achieved optimal operating conditions with 97.86% and 98.5% accuracy, demonstrating quick response times, low computational burden, and high customizability, suitable for practical online optimization and adaptable to other chemical industries. More recently, researchers have addressed the challenges of optimization in fed-batch biopharmaceutical processes by proposing an RL framework that incorporates human knowledge. Verified through the domain-specific simulator, the RL scheme enhances batch yields by 14% with minimal online computation time, demonstrating significant potential over existing methods. Additionally, a series of studies have explored using RL to generate real-time optimal decisions for digital twins. In summary, RL provides the benefit of training an optimal policy rather than optimizing actions at each time step. Once the optimal policy is identified, a cost-effective forward pass through the function can quickly produce an online solution, enabling faster decision-making.
Adaptive Optimization.  Adaptive optimization is a method that dynamically adjusts optimization parameters or strategies in response to real-time feedback and changing conditions to continually improve performance and handle uncertainties. However, traditional mathematical programming methods often involve relatively static models that struggle to repeatedly reconfigure or adjust parameters to handle complex and changing scenarios . By collecting real-time data and making corresponding adjustments, AI can better address these tasks. In the early days, research efforts focused on using traditional machine learning methods to achieve better adaptive optimization . Although the above results and methods can address optimization of process parameters to a certain extent, their dynamics are poor and they cannot well adapt to changing environments . Liu et al. proposes an AI-based adaptive optimization method for predicting surface roughness and adaptively adjusting process parameters in parts machining, addressing the limitations of traditional optimization methods in handling real-time and uncertain factors. By constructing a digital twin and utilizing a combination of Particle Swarm Optimization and Generalized Regression Neural Networks, the method enables real-time monitoring, prediction of tool wear and surface roughness, and adaptive optimization of cutting parameters, thereby improving both quality and efficiency in the machining process. More recently, Yang et al. present an adaptive optimization method combined with federated learning, which leverages the advantages of federated learning in handling heterogeneous data. This method improves convergence speed by over 60% and reduces traffic consumption by over 60%. By integrating various AI technologies, it is possible to more effectively overcome the limitations of traditional optimization methods in terms of dynamism, thereby enhancing adaptive optimization and improving this task for digital twins.
Adaptive Control
Adaptive control is a technique used to improve the efficiency or reduce the resource costs of physical systems by leveraging real-time simulation results and sensor data. Specifically, through enhancements of reinforcement learning, adaptive control plays a crucial role in various applications, including manufacturing equipment regulation , robotics , and autonomous driving . In this section, we focus on introducing the most representative AI technology utilized in control strategies - Reinforcement Learning (RL), and summarizing its advantages over traditional control techniques like Proportional-Integral-Derivative (PID) Control and Model Predictive Control.
Reinforced Learning and PID Control.  In the control tasks for digital twins, the most traditional and widely used controller is the PID controller. However, PID control struggles to handle high-dimensional, highly nonlinear, and time-varying systems. By integrating with Reinforcement Learning techniques, these challenges can be addressed to a certain extent. One approach involved using RL agents for parameter tuning of PID controllers. Early attempt used an incremental Q-Learning strategy to tune the PID controller in an online fashion. This algorithm dynamically grows a Q-value table in both the action space and state space direction through discretization techniques, in order to create a discrete yet accurate tuning model. Other methods have employed the Continuous Action RL Automata algorithm and a Radial Basis Function Actor-Critic network to tune a PID controller. Nevertheless, these methods do not fully resolve the linearity issues inherent to PID controllers. Furthermore, efforts have been made to entirely replace PID controllers with RL. One study successfully implemented the DDPG algorithm to develop an intelligent control strategy for the transient response of a variable geometry turbocharger system. Similarly, other researchers have proposed RL models as substitutes for PID controllers.
Reinforced Learning and Model Predictive Control.  Model Predictive Control (MPC) is also a widely used control technique, particularly suitable for control systems that require anticipatory actions. It is extensively employed in applications such as autonomous driving and robotic control. While MPC can handle complex, multivariable systems, it faces challenges with computational complexity and model accuracy. Integrating RL with MPC offers promising solutions to these issues. One approach involves using RL to improve the performance of MPC controllers. A digital twin-based adaptive controller has been proposed that integrates software-in-loop (SIL) and hardware-in-loop (HIL) simulations . This method allows for real-time optimization of the MPC parameters, enhancing its ability to handle nonlinear and time-varying systems. Another strategy employs RL to directly replace traditional MPC algorithms. Researchers have developed DQN policy models capable of regulating complex systems against realistic time-varying perturbations . These RL-based controllers can adapt to changing system dynamics without requiring explicit model updates, a significant advantage over traditional MPC. Digital twins play a crucial role in the development of RL-MPC systems. By creating accurate virtual representations of physical systems, digital twins provide safe environments for training and testing RL algorithms before deployment. For instance, an LSTM network has been used to capture the full dynamics of a control system, serving as a digital twin for RL training . The integration of RL and MPC has shown promising results for the control tasks of digital twins.
Towards Autonomous Management of Digital Twins
Cognitive Capabilities for Autonomous Management
Autonomous management relies on the cognitive layer of the digital twin, which allows it to understand human intentions and interpret the dynamic state of the system it manages. These cognitive capabilities form the foundation of intelligent control: the twin must first comprehend what needs to be achieved and perceive how the system is currently behaving. Large language models provide the mechanism for understanding and translating natural language commands into management operations, while foundation models enable perception across multiple data modalities. Together, they establish the semantic bridge between human intent, environmental understanding, and autonomous decision-making. It should be noted that current LLM- or diffusion-based world models do not guarantee physical fidelity or closed-loop stability, and their role in digital twins remains largely exploratory.
Natural Language Interaction with LLMs
Natural language interaction enables digital twins to understand and respond to human instructions in an intuitive way . Instead of relying on predefined scripts or manual configuration, operators can issue management goals directly using ordinary language, such as âreduce energy consumption without slowing productionâ or âcheck whether the network is operating normally.â The large language model interprets these instructions, extracts actionable entities, and translates them into formal goals that the twin can execute . This natural interface transforms digital twin management from a highly technical process into an accessible and collaborative activity .
From Language to Management Decisions.  The ability to transform human language into system-level actions represents a central cognitive advance for autonomous management. Large language models can parse free-form text into structured intents, identifying the relevant variables, constraints, and objectives involved in a task. They then map these structured goals to internal control modules or simulation tools through schema alignment and function invocation . For example, when a user requests âoptimize throughput while keeping temperature below 70°C,â the model converts the phrase into measurable objectives that define optimization targets and boundary conditions for the planning module. This process is reinforced through retrieval-augmented reasoning, where the model grounds its decisions in real-time system data and prior management records . By incorporating contextual information from sensor logs, historical performance data, and configuration files, the model avoids speculative reasoning and ensures that generated management actions are aligned with the current state of the physical system. As a result, language ceases to be a vague or ambiguous form of communication; it becomes a direct, interpretable interface between human expertise and machine execution. More importantly, this transformation also enables adaptability. When the system receives partially defined or conflicting goals, the model can request clarification, negotiate constraints, or infer missing details based on historical patterns. This bidirectional reasoning loop ensures that human intent is faithfully translated into precise operational instructions, supporting both efficiency and safety in autonomous control.
Conversational Management Interfaces.  Beyond one-time instructions, large language models enable ongoing dialogue between humans and digital twins. Through conversational interfaces, the system can provide summaries of its current state, report ongoing operations, and justify its planned decisions . For instance, after executing a maintenance optimization, the twin might respond: âCooling power has been reduced by 10%, energy consumption decreased by 12%, and no overheating detected.â Such transparency allows users to remain informed without micromanaging the process. Continuous conversation also allows for dynamic collaboration. Users can refine goals in natural language, such as âreduce the temperature faster but avoid overshoot,â and the system immediately adjusts its plan, re-evaluating trade-offs through its internal reasoning module . This adaptability turns management into a continuous negotiation between human and machine, where language serves as the medium for shared understanding. In addition, conversation provides a mechanism for accountability. The twin can explain why it made a particular decision, cite the data used, and quantify uncertainty in its predictions. These dialogue-based explanations not only enhance trust but also provide a foundation for regulatory transparency and auditability, which are essential in safety-critical domains such as manufacturing and infrastructure management.
Multimodal Perception with Foundation Models
A digital twinâs ability to manage autonomously depends not only on understanding instructions but also on perceiving its environment accurately. Foundation models extend perception beyond individual sensors by learning joint representations across images, signals, and text . This multimodal capability allows the twin to interpret complex operational contexts-identifying patterns, recognizing anomalies, and inferring hidden conditions that may not be apparent from a single data stream . In doing so, it bridges the gap between physical phenomena and digital awareness.
Multimodal Data Understanding.  Foundation models learn to interpret diverse data modalities by encoding them into a shared semantic space. They integrate information from cameras, vibration sensors, acoustic signals, thermal arrays, and textual logs to build a holistic understanding of the systemâs state . For example, a model may correlate a faint noise pattern with an abnormal vibration frequency, identifying a potential mechanical imbalance before traditional diagnostics detect it. This enables a form of perceptual intelligence that moves beyond static thresholds toward contextual reasoning. By abstracting multimodal signals into interpretable embeddings, the model can recognize system configurations, operating modes, and early indicators of degradation. This capability enhances situational awareness and provides a richer foundation for prediction and control. As environments evolve, the model continues to adapt through incremental fine-tuning, allowing perception to remain accurate even under new conditions or sensor configurations . Multimodal data understanding supports semantic consistency between the physical and digital layers. When the system observes discrepancies, such as when sensor readings suggest stability but video data reveals irregularities, it can flag these inconsistencies for review. This capacity to cross-validate signals across modalities strengthens reliability and prevents blind spots in autonomous management.
Multimodal Fusion for System Awareness.  Fusion mechanisms combine diverse streams of information into a unified situational representation that reflects the systemâs overall status . Through spatial and temporal alignment, the twin synthesizes localized signals into a coherent global view. For instance, data from distributed sensors across a power grid can be fused with satellite imagery and maintenance logs to infer both operational performance and environmental stress patterns. Such integrated awareness enables the twin to reason about interactions among components rather than treating each subsystem in isolation. This fusion-driven awareness also supports proactive behavior. The digital twin can detect emerging patterns, anticipate cascading effects, and adjust management strategies in advance. By tracking correlations across modalities, it learns causal relationships that link physical states to operational outcomes. The result is not just richer perception but an embodied understanding of the systemâs dynamics. Finally, maintaining awareness requires the twin to represent uncertainty transparently. Foundation models quantify prediction confidence and associate each inference with traceable evidence . This information is passed to the planning and decision modules, ensuring that high-risk judgments are handled conservatively or deferred to human oversight. In this way, multimodal fusion transforms raw data into actionable awareness, serving as the perceptual counterpart to the reasoning and planning processes that define autonomous management.
Enabling Autonomous Decision and Planning
While cognitive capabilities enable a digital twin to understand intentions and perceive the system state, autonomous management further requires the capacity to make decisions, plan actions, and adapt to changing conditions without explicit human intervention. This section discusses how agent-based reasoning and adaptive learning provide the mechanism for decision-making, and how self-optimization and closed-loop control complete the cycle of autonomous management. Together, these elements operationalize the MAPE-K paradigm, enabling digital twins to continuously monitor, analyze, plan, execute, and refine their management strategies.
Agent-based Reasoning and Adaptive Learning
Agent-based reasoning endows the digital twin with a modular structure capable of acting intelligently in complex environments . Each agent embodies autonomy, perception, reasoning, and learning, functioning as both a decision-maker and an executor of management tasks. Within a digital twin system, agents perceive environmental inputs, analyze contextual information, generate management plans, and execute actions through interactions with the underlying physical or simulated systems. Reinforcement and continual learning further allow these agents to refine strategies from experience and coordinate with others in multi-agent settings .
Agent Architecture for Autonomous Management.  The agent architecture forms the operational backbone of autonomous management. Typically organi2zed according to the MAPE-K loop, an agent continuously monitors system states, analyzes patterns, plans interventions, executes actions, and updates its knowledge base . This loop transforms management from reactive to proactive, allowing the digital twin to maintain stable operation even under uncertainty. The architecture usually comprises four layers: a perception layer that collects multimodal signals, an analysis layer that diagnoses conditions or predicts outcomes, a planning layer that formulates adaptive strategies, and an execution layer that interfaces with actuators or simulation modules . These layers interact bidirectionally, ensuring that every action is grounded in current observations and that outcomes feed back into learning and optimization. A practical example is the autonomous operation of industrial plants. Agents continuously assess performance metrics, detect bottlenecks, and coordinate actions such as adjusting temperature or resource allocation . When the system deviates from desired performance, the planning layer proposes new configurations and the execution layer implements them automatically. The knowledge component stores contextual rules, such as causal relationships between control variables and outcomes, enabling the system to reason about âwhyâ certain strategies work and reuse them in future scenarios. Furthermore, agent-based frameworks support hierarchical and distributed management. Local agents handle subsystem optimization, while higher-level agents coordinate global objectives, ensuring alignment between individual actions and system-wide performance. This organization enables scalability and robustness, as local failures can be compensated by neighboring agents without central control .
Learning from Management Experience.  Adaptive learning transforms the digital twin from a static control system into a self-improving entity. Through reinforcement learning (RL), agents learn management policies that maximize long-term rewards rather than immediate performance . Each experience, including success, failure, or anomaly, contributes to refining these policies, allowing the twin to anticipate the consequences of its actions in similar future contexts. Over time, this continuous improvement leads to resilient and efficient management strategies. For instance, consider an agent managing energy consumption in a data center. Initially, it may explore different cooling strategies through simulation. As it accumulates feedback on temperature stability, cost, and latency, it gradually learns an optimal control policy that balances energy efficiency and performance . This process mirrors human expertise acquisitionâlearning not only from positive outcomes but also from mistakes that inform better decisions. In addition to single-agent learning, collaborative learning across multiple digital twins enhances global intelligence. When experiences from one system are shared with another through federated or transfer learning, the collective knowledge base grows . This enables rapid adaptation to new environments without starting from scratch, promoting a form of âorganizational memoryâ for complex infrastructure networks. Ultimately, learning from management experience allows digital twins to evolve beyond pre-programmed behaviors, enabling self-adaptation in dynamic, uncertain, and data-rich environments.
Self-Optimization and Closed-Loop Control
Self-optimization is the culmination of autonomous management, where digital twins no longer rely on external commands but continuously refine their performance through closed-loop feedback . The system observes its own behavior, identifies inefficiencies, and implements corrective actions automatically. When combined with predictive and cognitive capabilities, closed-loop control transforms the twin into an autonomous entity capable of sustaining optimal performance with minimal supervision.
Autonomous Closed-Loop Management.  Closed-loop management completes the autonomous control cycle by continuously connecting perception, reasoning, and execution . In this paradigm, the digital twin monitors real-time data, detects deviations from expected performance, and triggers self-corrective actions. The MAPE-K loop becomes operational in a real-time context: Monitor to capture state data, Analyze to detect anomalies or predict outcomes, Plan to generate interventions, Execute to apply adjustments, and Knowledge to update future strategies . For example, in smart manufacturing, when vibration sensors indicate potential equipment fatigue, the digital twin predicts the failure horizon and autonomously schedules maintenance before a breakdown occurs . The system then evaluates the effectiveness of its intervention, learning from the outcome to improve future responses. This cycle of detection, action, and refinement ensures continuous adaptation to changing operational conditions. The advantage of autonomous closed-loop management lies in its capacity for sustained performance optimization. Unlike open-loop systems that depend on periodic calibration or manual adjustment, the closed loop enables continuous learning, model recalibration, and policy adaptation. Over time, the twin becomes more proficient in managing itself, bridging the gap between simulation-based optimization and real-world operational autonomy .
Human-AI Collaborative Management.  Despite advances in autonomy, complete independence from human oversight is rarely desirable or safe. Human-AI collaboration remains essential for balancing efficiency with accountability . Collaborative management frameworks define different autonomy levelsâfrom fully manual control to fully autonomous operationâallowing human intervention when uncertainty or risk exceeds predefined thresholds. At intermediate autonomy levels, the digital twin acts as an intelligent assistant that recommends actions, explains its reasoning, and executes tasks upon approval. This cooperative workflow enhances human decision-making by offloading complex computations while preserving transparency and trust . For instance, in power grid management, the twin may autonomously adjust voltage to stabilize supply but defer load-shedding decisions to human operators when the consequences are ethically or economically sensitive. The collaboration also extends to learning. Human experts can guide the twin by providing feedback on system responses or labeling exceptional cases for model retraining . This hybrid feedback accelerates learning convergence while ensuring that autonomous management aligns with human values and regulatory constraints. Ultimately, the goal of human-AI collaborative management is not to replace human judgment but to augment it-creating digital twins that are not only self-managing but also self-explaining and accountable within human-centered systems.
Applications
Healthcare System
The rapid adoption of electronic health records (EHRs), along with the emergence of digital and smart healthcare, has accelerated the integration of diverse technologies aimed at optimizing healthcare operations, improving patient outcomes, and reducing healthcare costs . Among these innovations, the digital twin technology plays a significant role by enabling the simulation of complex systems and the integration of virtual representations of real-world entities with AI to transform healthcare . In the following sections, we ourlined several major applications of digital twin (DT) technology across key healthcare domains, including clinical practice, clinical research, drug discovery, disease modeling, and precision medicine .
Clinical Workflow Optimization.  As hospitals transition from conventional practice to technology-driven information systems, digital twins are becoming essential for monitoring operations and managing resources . By integrating real-time and historical EHR data to simulate clinical resource usage and system-level processes, digital twins serve as adaptive virtual environments for operational decision-making. These models go beyond simple prediction by enabling healthcare administrators to test intervention scenarios and prescribe data-driven strategies to optimize staffing, bed utilization, and equipment deployment. . This simulation-based decision support ensures that resources are used optimally, reducing waste and operational costs. For example, Siemens Healthineers utilized digital twins in the Mater Private Hospital in 2018 to simulate MRI and CT workflows, identifying potential improvements and significantly enhancing the patient experience . Moreover, digital twins facilitate scenario planning and decision-making, allowing hospitals to prepare for various contingencies such as sudden influxes of patients or changes in healthcare policies . This leads to enhanced flexibility, resilience, and overall quality of care within the hospital environment. Karakra et al. introduced a digital twin model based on discrete event simulation to evaluate the efficiency of current healthcare delivery systems and assess the impact of service changes without disrupting daily operations . In 2019, Karakra et al. further developed HospiTâWin, a virtual hospital replica that enables healthcare providers to track patient pathways, monitor behaviors, and predict future outcomes .
style="width:85.0%" />
Clinical Trial.  In addition to streamlining healthcare workflows, the digital twin also significantly contributes to advancements in clinical research . It can offer transformative potential to augment the design and emulation of clinical trials, addressing persistent barriers such as absent control arms, small cohorts, and limited generalizability, and thereby advancing evidence generation in medical research . For example, Digital twins use advanced AI and machine learning models to create virtual patients, simulating counterfactual outcomes and allowing trials to maintain statistical power with fewer participants . This approach has potential to reduce the duration and cost of clinical trials by enabling concurrent simulations of multiple scenarios, thus reducing the need for extensive human trials and accelerating the development of new treatments . For example, in 2021, Qian et al. proposed SyncTwin, a method that learns a patient-specific time-constant representation from pre-treatment observations to create a digital twin that closely matches the target patient, enabling accurate counterfactual predictions . Similarly, in 2023, Das and Wang et al. introduced TWIN, a model that enhances clinical trials by creating personalized virtual models of patients using large language models . These digital twins simulate patient health trajectories and predict clinical outcomes based on real-world data, demonstrating the significant advancements in trial efficiency and precision that digital twin technology can offer.
Drug Discovery and Development.  Drug discovery involves identifying potential new medications, understanding their mechanisms, and optimizing their efficacy and safety before they can proceed to human testing . The drug discovery and development process using digital twin technology involves creating virtual models of biological systems to simulate disease mechanisms and drug interactions . It usually begins with the identification and validation of therapeutic targets by modeling cellular processes and interactions. Drug candidates are then screened through simulated interactions with these targets, allowing for optimization before preclinical testing . Digital twins in silico trials can help predict drug behavior and potential side effects, enhancing trial design and patient stratification in clinical stages . This method accelerates drug development, reduces costs, and improves success rates by providing a detailed understanding of drug behavior within a virtual biological system. In 2023, Moingeon highlights the utilization of digital twin in drug development against autoimmune disease, where digital twins of human cells are created using multi-omics data, including detailed information at the single-cell level . This data can be combined with machine learning to predict how cells will respond to different drug candidates, enabling rapid testing of billions of drug combinations. This not only accelerates target identification and optimization but also facilitates drug repurposing for novel therapeutic applications. As another example, Subramanian utilized digital twin to create a comprehensive virtual model that simulates liver function, disease progression, and treatment responses . This model integrates data from experimental studies and mathematical frameworks, allowing researchers to predict drug-induced liver injury and optimize drug candidates before clinical trials .
Disease Modeling, Diagnosis and Prognosis.  With its ability to create dynamic, patient-specific models, digital twin technology plays a critical role in advancing precision medicine by supporting early disease detection and trajectory forecasting . Digital twins can simulate a wide range of physiological and pathological processes by integrating data from multiple sources, such as medical imaging, electronic health records, genetic information, and real-time sensor data . They can replicate the functioning of organs, track the development of diseases, and predict how a disease will progress in a specific patient. The ability to simulate and analyze various medical scenarios without invasive procedures represents a significant advancement in medical technology and patient care . By running simulations, healthcare providers can test various scenarios, such as the impact of different medications or lifestyle changes, without any risk to the patient . This helps in early detection of potential health issues and proactive intervention. In 2024, Wang et al. applied digital twin technology to simulate the inhomogeneous optical properties of multi-core fiber, achieving unpixelated, high-resolution tumor imaging to enhance cancer diagnosis . Surian et al. utilized clinical and physiological biomarker data from various cohorts to develop a digital twin that simulates metabolic health profiles . This model accurately predicts the onset of chronic kidney disease (CKD) over a three-year period and effectively stratifies patients into different risk categories, facilitating early detection and improving management strategies for CKD. Similarly, Venkatapurapu et al. integrated a mechanistic model of Crohnâs disease (CD) and digital twin of each patient to predict the temporal progression of mucosal damage and healing, which provides detailed and chronological predictions of disease dynamics to support treatment decisions .
Personalized Treatment.  Personalized treatment involves tailoring medical interventions to a patientâs specific genetic, environmental, and behavioral characteristics, offering a more targeted and effective approach to care. . Similar to applications in disease diagnosis, digital twin technology enhances personalized treatment by creating virtual replicas of individual patients using data from various sources, but it focuses on comprehensive health insights and the detailed customization of treatment plans . By incorporating a patientâs genetic information, digital twins can simulate how different medications will interact with the individualâs unique genetic makeup, predicting potential side effects or adverse reactions before administering the treatment, thus increasing the efficacy of treatments and minimizing adverse effects . Moreover, digital twins support continuous monitoring and adjustment of treatment plans, ensuring that care remains optimal over time . For example, Martinez-Velazquez et al. constructed Cardio Twin, an architecture based on digital twin technology that integrates data from smartphone sensors, external devices, social networks, and medical records for real-time analysis . It optimizes personalized treatment for patients with cardiovascular diseases by analyzing patient-specific data to predict drug interactions and customize therapies, while autonomously managing lab data transmission and emergency service requests. Similarly, Wickramasinghe et al. adopted digital twin technology in personalized care for patients with uterine endometrial cancer, analyzing extensive patient data to develop predictive models for cancer progression and personalized treatment responses rather than relying on predefined theoretical frameworks . The clinician support system based on digital twins addresses the complexities and cognitive limitations associated with traditional cancer care by leveraging AI and machine learning to process large datasets, thereby enhancing the precision and efficacy of cancer treatment.
Biological System
The application of digital twin technology in biological systems represents a groundbreaking advancement with the potential to revolutionize various fields within life sciences . By creating precise virtual replicas of biological entities, digital twins enable researchers to simulate, analyze, and predict the behavior of complex biological processes in real-time . This innovative approach facilitates a deeper understanding of molecular interactions, cellular dynamics, and anatomical functions . It enhances research and development, improves diagnostic techniques, optimizes experimental and therapeutic interventions, and streamlines production processes in biomanufacturing .
Molecular Research.  Application of digital twin technology can expand to various levels in biological systems . At the molecular level, digital twins can simulate the interactions between proteins, nucleic acids, lipids, and other biomolecules with high precision . These simulations encompass detailed molecular dynamics, including protein-protein interactions, DNA-RNA transcription and translation processes, lipid bilayer formation and behavior, and complex molecular assemblies . Consequently, digital twins provide comprehensive insights into complex biochemical pathways and molecular mechanisms, facilitating the development of targeted therapies, personalized medicine, and biotechnological innovations . For example, Hengelbrock et al. utilized digital twins for virtual replicas of the physical mRNA transcription process to simulate and optimize production conditions . By determining key kinetic parameters and using a plug flow reactor for high-throughput screening, the researchers significantly enhanced mRNA yield and reduced impurities like truncated mRNA. This approach enabled the efficient production of 20 vaccine candidates in a short time, a tenfold increase in productivity. The integration of process analytical technologies and digital twin within a Biopharma 4.0 framework facilitated continuous and automated production, ensuring a scalable and resilient supply of mRNA therapeutics. Similarly, Silva et al. leveraged digital twins to optimize the chromatographic process used for the purification of monoclonal antibodies (mAbs) . The digital twin framework facilitated the screening of Cation-Exchange (CEX) resins, model calibration, and the prediction of chromatographic behavior under different conditions, thereby streamlining process development, reducing experimental workload, and accelerating timelines.
Cell Manufacturing and Metabolic Pathway.  Digital twins at the cellular level create precise, real-time virtual models of biological processes, integrating data from various sources to simulate and optimize conditions for cell growth, production, and metabolic activity . These digital replicas enable researchers and manufacturers to monitor, predict, and adjust parameters in real-time, significantly enhancing efficiency and ensuring consistent product quality . In cell manufacturing, digital twins optimize bioreactor conditions by fine-tuning parameters such as temperature, pH, and oxygen levels, managing nutrient supply to maintain optimal growth conditions, and predicting product quality through continuous monitoring and feedback loops . For metabolic pathways, they simulate intricate biochemical processes, including enzyme activities, metabolic fluxes, and the effects of genetic modifications, thereby improving the design and productivity of cell-based processes . For instance, Cheng et al. incorporated a digital twin model with Biological Systems-of-Systems (Bio-SoS) to simulate interactions in sub-models like single-cell and metabolic shift models. This approach enables real-time adjustments of cultivation conditions and optimal nutrient management, ensuring consistent and high-quality cell culture outcomes . By simulating the interactions at various scales, from molecular to macroscopic levels, this digital twin model provides a comprehensive understanding of the cell culture process, facilitating more efficient and flexible manufacturing. Likewise, Helgers et al. employed a digital twin to enhance CHO cell-based antibody production using a dynamic metabolic model that simulates central metabolic pathways, such as glycolysis, the TCA cycle, and amino acid metabolisms . This model incorporates detailed reaction kinetics and feedback mechanisms, allowing for precise control of metabolic activities and cell growth. The digital twin demonstrated high accuracy and precision, enabling the optimization of yield and product quality in monoclonal antibody manufacturing. By integrating real-time data and predictive modeling, this digital twin supports continuous process improvement and robust process control, ensuring high productivity and compliance with regulatory standards.
Single Anatomical Structures.  Single anatomical structures are individual components of the body with specific forms and functions, typically recognizable as distinct entities, such as organs, bones, and muscles . Digital twins for these structures provide detailed, accurate models that enhance research and education . They enable pre-surgical planning by allowing surgeons to simulate complex procedures, thereby reducing risks and improving outcomes . Additionally, digital twins facilitate the simulation and testing of medical devices, ensuring safety and efficacy before application . These virtual models are created using high-resolution imaging techniques like MRI, CT scans, and ultrasound, integrating anatomical and physiological data to enable advanced simulations of real-life scenarios through deep learning techniques . For example, Shu et al. introduced Twin-S, a digital twin model designed for skull base surgery, which simulates, monitors, and continuously updates all essential aspects of the procedure in real-time to replicate real-world conditions . The model utilizes high-precision optical tracking and real-time simulation to create detailed virtual models of surgical tools, patient anatomy, and surgical cameras. By integrating data from sources like CT scans and employing calibration routines, the system ensures accurate representation and updates the virtual model at a frame rate of 28 FPS. Evaluation of Twin-S shows an average drilling simulation error of 1.39mm, demonstrating its accuracy and potential to improve surgical planning and outcomes.
One particularly notable advancement in this field is the development of digital twin models for the brain . The Digital Twin Brain (DTB) technology integrates multimodal neuroimaging data, genomic data, behavioral data, and cognitive assessments to construct personalized brain models that accurately simulate anatomical structures, functional connectivity, and dynamic changes . This virtual replica holds immense potential for neuroscience research, clinical practice, and brain-computer interfaces . In 2024, Park, Wang, Guan, and colleagues unveiled an integrated platform for multiscale molecular imaging and phenotyping of the human brain, creating a 3D atlas at subcellular resolution . This innovative approach simultaneously maps brain-wide structures and captures high-dimensional features. The platform includes “Megatome,” a device that finely slices intact human brain hemispheres without causing damage, and “mELAST,” which makes each brain slice clear, flexible, durable, expandable, and quickly, evenly, and repeatedly labelable. Additionally, the “UNSLICE” computational system seamlessly reunifies the slabs, reconstructing each hemisphere in full 3D with precise alignment of individual blood vessels and neural axons. This platform is expected to enable comprehensive analysis of numerous human and animal brains, enhancing our understanding of interspecies similarities, population differences, and disease-specific characteristics. It also facilitates the mapping of single-neuron projectomes integrated with molecular expression profiles, uncovering the organizational principles of neural circuitry and their alterations in diseases, thus advancing our understanding of disease mechanisms.
Multi-Scale Biological Systems.  Digital twins at the multi-scale level aim to capture the interplay between biological processes that unfold across molecular, cellular, tissue, and organ systems . Unlike single-scale twins, which focus on local dynamics, multi-scale digital twins need to reconcile differences in spatial resolution, temporal dynamics, and data modalities across layers . This creates fundamental modeling challenges: aligning scales with inconsistent data densities, integrating mechanistic models with statistical surrogates, and updating the system in real time while maintaining internal consistency . Additionally, many biological interactions are bidirectional: molecular perturbations may lead to organ-level changes, but systemic states such as inflammation or hormonal feedback can also rewire cellular behaviors . To address this complexity, multi-scale twins often rely on modular architectures, where distinct sub-models (e.g., gene regulatory networks, metabolic flux simulators, biomechanical solvers) are coupled through interface layers . These models support high-fidelity simulation of emergent physiology and enable hypothesis testing for diseases whose behavior cannot be understood at a single biological level, such as cancer progression, organ failure, or complex neurological disorders .
Multiâscale digital twins put these frameworks into practice by enabling integrative modeling of complex disease systems that defy single-level explanation. For example, in oncology, TumorTwin constructs patientâspecific breast cancer twins by linking genomic profiles, histopathology features, tumor microenvironment data, and longitudinal imaging, allowing simulation of tumor evolution and therapy response across molecular, cellular, and anatomical scales . In organ failure, Gallo et al. developed a liver twin that integrates hepatocyteâlevel metabolism, zonated lobular blood flow, and wholeâorgan perfusion, enabling prediction of spatially localized drug-induced liver injury and systemic decompensation . For neurological disorders, virtual brain twins for stimulation in epilepsy established personalized brain digital twins by combining patientâspecific structural MRI/DTI, networkâscale neural population models, and EEG/SEEG data . This multiâscale architecture captures how stimulation perturbs local excitability and triggers seizure propagation throughout the brain, supporting individualized estimation of epileptogenic zones and in silico testing of neurostimulation strategies. Across these domains, multiâscale digital twins serve as unified platforms that integrate data and mechanisms from molecular or cellular levels to organ and system scales, enabling mechanistic insight and actionable simulations.
Aerospace
The aerospace sector integrates multiple disciplines of engineering, data analytics, and systems control to support the development and operation of aircraft and spacecraft . Because of its complexity and strict safety requirements, the industry has rapidly adopted digital twin technology as a foundation for modernization . A digital twin creates a virtual model that mirrors the physical system, enabling continuous monitoring, simulation, and predictive analysis . When integrated with artificial intelligence and physics-based modeling, these systems allow adaptive design, predictive maintenance, and autonomous decision-making, marking a major transformation in aerospace engineering .
Aircraft Design.  At the design stage, digital twins support virtual prototyping and early verification of aerostructures . By linking computer-aided engineering models with computational fluid dynamics and finite element analysis, engineers can explore aerodynamic performance and structural integrity before physical testing . Li et al. introduced a digital twin framework for composite part fabrication that synchronizes process data with simulation feedback to minimize deviation. In Europe, the Clean Sky 2 program used digital twins to model airframe deformation under load and refine lightweight materials, achieving measurable improvements in fuel efficiency . These developments demonstrate how digital twins shorten design cycles and increase accuracy in aircraft development.
Maintenance and Health Monitoring.  Maintenance has become one of the most mature and widely adopted applications of digital twins in aerospace . By integrating real-time sensor data with high-fidelity physical models, engineers can estimate the remaining useful life of key components and identify early signs of fatigue . The U.S. Air Force developed the Airframe Digital Twin project to simulate cumulative damage and schedule inspections more effectively . Siemens has also applied digital twin systems to engine monitoring, combining vibration and temperature data to predict performance degradation and prevent unplanned failures . With advances in AI-based diagnostics, these predictive models are helping the aerospace industry transition from reactive maintenance to proactive asset management.
Flight Simulation and Mission Planning.  Digital twins are also reshaping flight simulation by creating dynamic digital environments for training and mission testing . Through integrated aerodynamic and structural modeling, engineers can evaluate flight stability and control under varying atmospheric and operational conditions . NASA incorporated digital twin environments into its simulation platforms to test autonomous flight algorithms and fault recovery procedures before live deployment . These virtual testbeds have reduced the reliance on costly physical experiments and improved pilot decision support, particularly in missions that require real-time adaptability.
Space Missions and Satellite Systems.  In space operations, digital twins assist mission controllers in managing spacecraft health and predicting system behavior in orbit . NASAâs Artemis program applies digital twins to simulate propulsion, communication, and thermal subsystems, allowing mission planners to evaluate performance under extreme conditions . For satellite constellations, twin-based simulators reproduce orbital dynamics and sensor interactions, helping teams predict fuel usage and detect anomalies before they escalate . The European Space Agency has used similar architectures for ground control validation, improving reliability and reducing mission delays. These applications demonstrate the growing importance of digital twins in maintaining the safety and longevity of space assets.
Autonomous Space Robotics.  Digital twin technology is equally transformative for robotic systems used in orbit and planetary exploration . By constructing high-fidelity models of robotic manipulators and rovers, engineers can test control strategies and failure responses under microgravity and harsh terrain conditions . NASA employs digital twin frameworks to monitor robotic arms on the International Space Station, predicting mechanical wear and optimizing motion planning . In research by Grinshpun and Rossmann, twin-based virtual testing enabled autonomous robots to rehearse debris removal and satellite servicing tasks, ensuring operational safety before deployment . These developments highlight how digital twins are enabling more resilient, self-learning robotic systems that can operate independently in remote environments.
In summary, digital twin technology is redefining aerospace engineering by fusing data, physical modeling, and intelligent analytics into continuous feedback systems. From early aircraft design to autonomous space missions, these technologies improve precision, reliability, and efficiency throughout the aerospace lifecycle.
Smart City
The rapid pace of urbanization, coupled with the growing demand for livelihoods among urban dwellers and advancements in technology, has significantly accelerated the development of smart cities . Smart cities integrate a range of innovative technologies, including digital twins, the Internet of Things (IoT), blockchain, and AI, to deliver smart services and enhance the quality of life for local residents . Among these technologies, digital twins offer virtual replicas of real urban environments, providing a robust platform for enhancing smart city capabilities . Additionally, AI further strengthens the data-driven foundations of smart cities by improving data accuracy, predictive analytics, and decision-making . This segment explores the applications of digital twin and AI technologies within the context of smart cities.
Building Efficiency.  Buildings account for the majority of energy consumption in cities , making them a critical area of attention for the development of smart city initiatives . In the context of urban building energy systems, digital twins significantly enhance the capabilities of smart metering infrastructure. This infrastructure facilitates the recording of electricity usage at granular levels, with meter readings taken at intervals of less than one hour . Beyond electricity metering, building-level sensing infrastructures increasingly integrate Internet of Things sensors to capture environmental and operational variables such as occupancy presence and indoor temperature, which are critical drivers of building energy demand . Recent studies highlight that such IoT-enabled sensing infrastructures constitute the primary source of large-scale urban and building data used in data-driven energy management systems . These heterogeneous sensor streams form the data foundation for data-driven building energy modeling and control. When integrated with machine learning techniques, this rich energy data can support a wide range of applications, including energy load analysis, forecasting, management, and real-time assessments of energy consumption . For instance, decision tree and support vector machine (SVM) classifiers can be employed to detect anomalous consumption patterns in real time . Additionally, a k-nearest neighbors classifier can be utilized to analyze the energy behaviors of occupants in commercial buildings . For building energy prediction, Fan et al. leveraged deep learning techniques for data augmentation, significantly improving the accuracy of short-term energy forecasts for buildings .
Microgrids.  Microgrids and smart grids are among the most important and well-established application domains of digital twin technology. Unlike building-level energy systems, microgrids operate at the intersection of local energy generation, distributed loads, and the public power grid, extending beyond the scope of building energy efficiency alone. A typical microgrid integrates distributed energy resources, interconnected loads, energy storage systems, and control mechanisms, enabling flexible operation under both grid-connected and islanded modes. Within the digital twin framework, microgrids have been extensively investigated for their potential to enhance system-level intelligence, robustness, and operational resilience. Existing studies have focused on forecasting tasks, including renewable generation and load prediction , as well as system management and real-time monitoring . Digital twin models have also been applied to fault detection and predictive maintenance, facilitating the early identification of abnormal behaviors and component failures . In addition, cybersecurity and system protection have emerged as critical research topics, particularly given the increasingly tight cyberâphysical coupling in microgrids . A prominent research direction involves using digital twins to proactively enhance grid robustness against multifaceted uncertainties. Advanced frameworks employ hybrid stochastic-robust optimization methods to determine optimal schedules for both normal and resilient operation, effectively modeling uncertainties related to grid costs, renewable generation, and loads . Beyond standalone resilience, digital twins are crucial for deepening the operational and supportive integration of microgrids with the main public grid. This involves not only islanding during outages but also providing active support to the main grid under normal conditions. For example, digital twin-enabled coordinated control strategies can orchestrate distributed energy resources such as wind and solar to provide rapid frequency support to the main grid, a critical service as renewable penetration increases . At the local distribution level, digital twins play an indispensable role in coordinating complex, modern energy consumption, particularly within building clusters and electric vehicle charging stations. The spatiotemporal flexibility of electric vehicles, including their ability to move and shift charging times, presents both a challenge and a unique resource for local energy management. Recent research proposes digital twin-based strategies for coordinating electric vehicle charging and discharging across multiple temporary microgrids formed after a blackout . Beyond these aspects, Sun et al. emphasize the role of digital twins in improving microgrid robustness and resilience against various sources of uncertainty, including renewable intermittency, load fluctuations, and grid disturbances . Lasseter et al. highlight the growing research attention toward the integration of microgrids with the public grid, especially when considering localized energy consumption scenarios such as building clusters and electric vehicle charging stations . Furthermore, Palensky et al. demonstrate that digital twin-enabled co-simulation and control strategies provide a promising approach for coordinating energy flows across multiple subsystems and operational scales .
Urban Planning.  Infrastructureâincluding buildings, bridges, roads, railways, and associated pipeline networksâplays a critical role in the development of smart cities. However, planning, constructing, and maintaining such infrastructure presents significant challenges due to high costs, extensive resource requirements, long project timelines, and the unique characteristics of each project . Digital twin technology offers a promising approach to address these challenges. A fundamental aspect of digital twin technology is the representation of infrastructure. Digital twins of cities can be organized into multiple layers , one of which is the geometric layer . This layer captures the morphology and physical characteristics of infrastructure in 3D . For example, 3D digital twin modeling software like CityEngine can be used to manually reconstruct urban infrastructure, as demonstrated in Matera, Italy . The accuracy of these models can be further enhanced using AI and data augmentation techniques , while deep learning methods enable the generation of complex, high-dimensional digital representations of urban infrastructure . Beyond modeling, digital twin technology supports infrastructure design through simulation and analytical testing. By relying on virtual models, designers can perform 3D design iterations with greater precision and receive more accurate feedback compared to traditional methods . This virtual approach extends to fabrication and installation, where digital twins combined with virtual reality allow for systematic data storage, precise positional coding, and accurate measurement of components . Integrating data-driven digital twin models with intelligent systems further enables real-time predictive maintenance, diagnostics, and informed decision-making . Machine learning techniques, for instance, can support Structural Health Monitoring (SHM) of civil infrastructure . These methods allow for independent generation of new data samples, enhancing infrastructure safety by predicating potential failures.
Public Safety and Environment.  Digital twin technology plays an increasingly important role in enhancing public safety and environmental governance by offering a dynamic, data-driven framework for simulating, monitoring, and managing complex urban systems. In this context, its primary applications encompass emergency management, environmental risk assessment, and the protection of critical infrastructure. In the field of emergency management, digital twins serve as effective tools for disaster preparedness, real-time response, and post-event recovery. By integrating real-time data streams from IoT sensors with geospatial information systems and historical incident records, digital twins generate a continuously updated virtual representation of a cityâs physical state. This living model enables authorities to visualize the evolution of natural hazards, such as floods, earthquakes, and wildfires, in real time, assess their potential impacts on populations and infrastructure, and evaluate alternative evacuation plans or resource allocation strategies through scenario-based simulations . During flood events, for instance, digital twins can model hydrodynamic processes, predict inundation extents, and identify vulnerable critical assets, thereby supporting early warning dissemination and targeted emergency interventions . Beyond reactive disaster response, digital twins also support proactive environmental monitoring and public health protection. By assimilating data from air quality sensors, traffic flows, and meteorological observations, digital twins can model and forecast urban air pollution dynamics, enabling the identification of pollution hotspots and the evaluation of mitigation measures such as traffic restrictions or emission control policies . In a similar manner, digital twins can be applied to monitor the spatial and temporal distribution of noise pollution and assess its effects on urban livability and community well-being. Digital twin technology further contributes to public safety through the continuous monitoring and protection of critical infrastructure. As demonstrated by Khan et al., machine learning techniques embedded within digital twin frameworks enable the early detection of structural anomalies or performance degradation in assets such as bridges, tunnels, and power grids . By continuously comparing real-time sensor data from physical structures with the expected behavior predicted by their digital counterparts, potential failures can be anticipated before they occur. This enables predictive maintenance strategies and reduces the likelihood of catastrophic incidents. Such capabilities are particularly valuable for advanced composite materials used in aerospace and transportation systems, where digital twins facilitate structural health monitoring (SHM) to ensure long-term integrity and operational safety .
Mobility and Transportation
The growing trend towards connectivity and automation in the transportation sector is increasingly leveraging advanced technologies to optimize traffic management and improve overall system efficiency . A particularly promising technology is the digital twin, which can be employed to create dynamic and real-time models of various elements within the transportation ecosystem. This includes the ability to monitor and simulate traffic flow, evaluate vehicle performance, and assess road conditions, among other critical factors that collectively impact the effectiveness and safety of transportation systems . By providing a comprehensive and continuously updated virtual representation of the physical world, digital twin technology stands as an essential tool at the forefront of mobility and transportation innovation. Given the rapid advancements in Connected and Automated Vehicles (CAV) and the Internet of Vehicles (IoV) technologies, the integration of digital twins into traffic management platforms is becoming increasingly feasible and advantageous. This integration promises to further revolutionize mobility and transportation systems by enabling safer and more efficient traffic management solutions .
Connected and Automated Vehicles.  The rise of CAV technology introduces another platform to implement digital twins beyond traditional autonomy. Since the level of automation and connectivity within our vehicles has greatly improved, these equipped vehicles can not only sense their surroundings using their onboard perception sensors (e.g., camera, LiDAR, radar) but also âtalkâ with other agents such as vehicles, infrastructure through vehicle-to-everything (V2X) communications . Various studies leverage cloud computing to empower digital twins to serve for CAVs. For example, Wang et al. proposed a mobility digital twin framework that empowers CAVs with various micro-services based on a device-edge-cloud architecture . Alam and Saddik designed a digital twin framework reference model for the cloud-based CPS, where a telematics-based driving assistance application was proposed for the vehicular CPS with three parts: 1) computation, 2) control, and 3) sensors and services fusion . Kumar et al. developed a digital twin-centric approach with machine learning, edge computing, 5G communication, and data lake, aiming for driver intention prediction and traffic congestion avoidance . Wang et al. proposed a vehicle-to-cloud paradigm for an advanced driver-assistance system (ADAS) of CAVs , and this paradigm was further experimented with by Liao et al. in a cooperative ramp merging scenario . The challenge of properly visualizing the digital twin information received from the cloud was studied by Liu et al., where a sensor fusion method that combines onboard camera data was proposed to facilitate the decision-making of CAVs . Gao et al. proposed a multi-tiered Carla-SUMO-AirSim co-simulation approach to bridge ground and air V2X collaboration, enabling broader embodiment for CAV research . Testing and evaluation are essential for the progress and implementation of CAVs. Digital twin technology is particularly useful for facilitating closed-facility testing by integrating virtual components generated by computer systems with real-world road conditions . Fully testing automated driving systems poses significant challenges and requires extensive testing that cannot be accomplished without simulation support. As a result, recent research has concentrated on creating simulation frameworks for testing . The digital twin framework proposed by Ge et al. identifies three testing levels: entirely virtual, based on actual sensor data, and vehicle-based . A similar approach involves using digital twin technology within specific frameworks to record vehicle responses in various simulated environments, which helps generate a large dataset for training and testing autonomous vehicle control systems, thus establishing a solid foundation for accurate system development . Digital twins are also used to create statistical models that forecast a vehicleâs future movements based on historical data . Additionally, researchers have applied digital twin technology to dynamically compute motion parameters, improving vehicle trajectory planning . The modeling of human behaviors on CAVs has also been advanced with digital twin technology, where the Driver Digital Twin (DDT) concept was first proposed by Chen et al. to simulate human driver behavior on CAVs . This concept has been further researched by various studies, which involve digitizing human drivers to link current autonomous driving systems with fully digital systems and predicting driversâ future decision-making process with machine learning approaches . This development contributes to the creation of a comprehensive Human-Cyber-Physical System (H-CPS) that integrates human driving behaviors .
Internet of Vehicles.  IoV is a sophisticated network of vehicles outfitted with sensors, software, and technology, all of which follow standardized protocols to connect and share data over the Internet . This network includes not only the connections between vehicles but also the links between vehicles and various road infrastructures and the cloud . By exchanging traffic information, IoV systems can collaboratively optimize vehicle movements and traffic control, leading to a significant improvement in overall traffic efficiency . As a result, IoV is seen as a crucial factor in the future of autonomous driving, as well as connected, electrified, and shared mobility . Digital twin technology is widely utilized in IoV for resource allocation, sharing, and traffic forecasting. For example, a method for real-time traffic data prediction using digital twins has been developed, which relies on monitoring traffic flow and speed data transmitted via 5G through IoV sensors, greatly enhancing the systemâs accuracy and response time . Experiments were carried out using traffic data collected in Nanjing, China . Additionally, W. Sun et al. have explored a digital twin model aimed at dynamic resource allocation in aerial-assisted IoV networks, allowing for coordinated resource scheduling and distribution . In terms of offloading strategies within IoV, a study has proposed a digital twin network framework that maintains digital twins in cyberspace, facilitating the synchronization of real-world vehicle activities . An innovative IoV framework has also been introduced, utilizing digital twins to create a digital representation of the IoV environment . This enables real-time updates to a vehicleâs driving route based on current data, significantly improving navigation and operational efficiency. Moreover, a digital twin design that incorporates consortium blockchain technology has been developed, focusing on remote resource sharing within the IoV framework to effectively track and safeguard resources . In aerial IoV networks, the RADiT framework has been introduced as a digital twinâdriven resource allocation model that leverages real-time network state representations to enable fast and efficient resource sharing, thereby significantly enhancing network throughput, latency performance, and overall connectivity . As a prime example of a fully connected scenario, IoV enables intelligent vehicle operations through the use of artificial intelligence (AI). The future of vehicular networks will require a wide range of services that demand considerable computing resources. To tackle the resource shortage, neural networks are utilized to optimize the use of excess computing capacity. Federated Learning (FL) has also been applied extensively in the IoV sector . For instance, a novel asynchronous FL method has been created to ensure secure and efficient data sharing within IoV networks . Additionally, I. Ullah et al. introduced a blockchain-supported FL algorithm aimed at enhancing knowledge sharing in IoV networks, promoting collaborative learning and data integration . To address the challenge of efficiently recognizing license plates in 5G-enabled IoV environments, a new FL model has been designed to improve recognition accuracy and processing speed . Furthermore, FL-based collaborative positioning technology has been demonstrated in IoV networks to support autonomous driving and collision avoidance, thereby enhancing the safety and efficiency of vehicular movements .
Traffic Flow.  The application of digital twin technology in traffic flow presents considerable opportunities for enhancing mobility and transportation systems. Digital twins can effectively visualize traffic patterns within urban environments, providing a detailed and dynamic representation of traffic flow . VISSIM, a microscopic traffic flow simulation software, can be integrated into a digital twin framework for smart traffic corridors by utilizing real-world data, allowing for highly accurate simulations and analyses . A digital twin focused on mobility management has been created, leveraging a cloud-based microservices architecture to manage and optimize traffic flow in a scalable and flexible manner . Additionally, a radar-camera fusion approach has been proposed to develop a digital twin for specific sections of roadway, enhancing the precision and reliability of traffic monitoring and management . A traffic flow prediction model based on digital twin architecture and neural networks has also been established, enabling the accurate prediction of inflow and outflow at nodes within the Beijing Subway network . Another line of approaches for predicting network traffic and behavior of traffic participants involves the use of reinforcement learning or generative models, which have demonstrated effectiveness in providing long-term forecasts . Once traffic flow is predicted and visualized, the functionality of digital twins can be further enhanced by integrating artificial intelligence to classify traffic congestion. For instance, neural network models can leverage extensive datasets to improve the precision of traffic congestion classification, helping to identify and respond to congestion issues in real time . Moreover, digital twins can provide actionable guidance to drivers based on predicted traffic conditions. For example, advisory messages generated by a digital twin at an intersection can instruct drivers to adjust their speeds, promoting smoother vehicle movement without the need for frequent stops. This approach significantly reduces travel time and energy consumption, thereby improving overall mobility and reducing environmental impact . Furthermore, digital twins can play a critical role in optimizing traffic signal timings. By analyzing predicted traffic flow data, digital twins can dynamically adjust signal timings in a smart connected corridor testbed. Such optimizations have been shown to achieve a 20.81% reduction in travel times compared to traditional actuated traffic control systems, underscoring the potential of digital twins to enhance traffic efficiency and reduce congestion .
Smart Manufacturing
We are currently in the midst of the Fourth Industrial Revolution, also known as Industry 4.0. Industry 4.0 can be defined as the integration of intelligent digital technologies into manufacturing and industrial processes . With the maturity and application of new-generation information technologies, the advancement of Industry 4.0 is in full swing . A key component of Industry 4.0 is smart manufacturing , which represents the primary application of âmanufacturing intelligence" across the production and supply chain . Smart manufacturing refers to a new manufacturing paradigm where manufacturing machines are fully connected through wireless networks, monitored by sensors, and controlled by advanced computational intelligence to improve product quality, system productivity, and sustainability while reducing costs . This area has garnered widespread attention and experimentation . The implementation of smart manufacturing is driven by digital twins, which abstract physical entities in factories into their digital forms within cyberspace . This abstraction enables the monitoring, control, diagnosis, and prediction of the states of these entities . In addition to digital twin technology, smart manufacturing leverages a set of technologies that include industrial IoT networks, AI, Big Data, robotics, and automation to enhance system efficiency and improve outcomes . Among these, AI is a critical technology for digital twins in smart manufacturing. AI enables data processing and real-time prediction of manufacturing processes and component performance, thereby optimizing the performance of digital twins and improving product design and manufacturing efficiency . Here are some applications that illustrate the use of digital twins and AI to achieve smart manufacturing in the background of Industry 4.0.
Manufacture Visualization.  Visualizations, as an important component of manufacturing, can effectively transform vast amounts of information into knowledge and insights, thus facilitating system control for staff . In smart manufacturing, digital twins primarily drive graphical visualization, 3D interactive visualization, and augmented reality (AR) among the visualization technologies. Graphical visualization refers to the real-time mapping of machine status through visual representations, enabling operators to remotely view manufacturing data and better adjust systems. For instance, dashboards can display machine health status, production efficiency, and order scheduling . Additionally, Tong et al. presented HMIs and applications for visualizing and analyzing machining trajectories, machining statuses, and energy consumption . 3D interactive visualization involves 3D simulations of manufacturing equipment, processes, and products. It provides operators, technicians, and process planners with a more intuitive perspective to quickly identify issues and determine root causes. 3D interactive visualization can create virtual models of manufacturing equipment based on physical and kinematic models. For example, a virtual model of a grinding machine based on 3D, physical, and kinematic models can accurately reflect the equipmentâs real-time status and processes . Moreover, 3D interactive visualization enables remote access to physical systems, permitting experts to provide remote support and problem-solving without needing to be on-site . Manufacturers can gain a comprehensive view of manufacturing process data and analysis by combining 3D interactive visualization with other methods, such as dashboard monitoring. For instance, Zhao et al. integrated dashboards and 3D interactive visualization for the programmable logic controller (PLC) of a milling machine . The vision of Industry 4.0 is to construct cyber-physical production systems (CPPS) that seamlessly connect the physical and digital worlds, making manufacturing increasingly intelligent . AR applications enable real-time access to the vast data generated by CPPS , aligning with the needs of Industry 4.0 and smart manufacturing. Compared to 3D interactive visualization, AR can overlay virtual information onto the real world, further enhancing human-machine interaction. Operators can wear AR devices during inspections, using actual production scenes as the background to visualize the operating data of the equipment . It is worth saying that AR devices can be used to display possible defects on the product being inspected .
Production.  Production monitoring is a crucial aspect of manufacturing. In smart manufacturing, monitoring includes the condition monitoring of machines as well as the quality monitoring of products . Manufacturing machines often experience failures as a result of degradation or abnormal operating conditions, leading to increased operational costs, reduced productivity, higher rates of defective parts, and even unexpected downtime. Hence, the implementation of condition monitoring is imperative. This involves monitoring and tracking machine status, detecting early defects, diagnosing the root causes of failures, and integrating this valuable information into manufacturing production and control processes . With the help of digital twins, condition monitoring of industrial equipment can be achieved. A digital twin architecture was suggested and put into practice for a pneumatic robotic gripper to identify anomalies like pneumatic cylinder leaks and bearing malfunctions . Miao et al. demonstrated a digital twin framework using multidimensional time series data for anomaly prediction and equipment state monitoring for computer numerical control(CNC) machines . Increasingly, deep learning techniques are being extensively researched for condition monitoring, offering higher accuracy and timeliness . CNNs integrate feature learning and defect diagnosis into a single model and have been applied in many areas, such as bearings , gearboxes , wind generators , and rotors . In addition, deep Belief Networks (DBNs) have been investigated for fault diagnosis of aircraft engines , chemical processes , reciprocating compressors , rolling element bearings , high-speed trains , and wind turbines . Quality monitoring involves monitoring the quality of products and then identifying potential product defects to improve overall product quality. Compared to traditional quality monitoring, which is time-consuming, labor-intensive, and unable to detect subtle defects, digital twins offer a more efficient quality monitoring method. Using manufacturing and sensor data, a digital twin-based predictive model for surface roughness quality in machine tools was developed . To tackle optical quality control challenges in additive manufacturing, multiple digital twins were developed to streamline the inspection process . Deep learning, especially CNNs, has been applied to various textures or hard-to-detect defect cases. CNNs, initially designed for image analysis, are particularly suited for automatic defect identification in surface integration inspection. Max-pooling CNNs perform feature extraction directly from pixel representations of steel defect images , facilitating automatic inspection of dirt, scratches, burrs, and wear on surface parts .
Optimization in industrial manufacturing refers to improving performance and efficiency while meeting specific objectives within manufacturing operations. Digital twin optimizes manufacturing processes by providing accurate, real-time data to enhance performance, reduce waste, and increase sustainability through virtual commissioning and parameter optimization  . Virtual commissioning involves using the virtual entities created by digital twins to optimize and debug processes, replacing some physical operations. This helps reduce operator fatigue and safety risks. For example, simulation tools can be used to retrofit traditional machine tools . In industrial manufacturing, continuous trial and error are required to find parameters that yield the highest quality, efficiency, and benefits, which is known as parameter optimization . However, traditional physical methods of trial and error are often insufficient, costly, and time-consuming. Digital twins offer a significant advantage by simulating virtual scenarios and predicting processes, allowing for low-cost testing of various parameters to find the optimal values. Consequently, many applications of digital twins focus on parameter optimization. For instance, a digital twin system of a cutting machine tool can be developed to optimize machining dynamics and estimate and compensate for contour errors . Furthermore, Balderas et al. applied ant colony optimization for manufacturing hole patterns on printed circuit boards using minimal trajectory and tool change time . Moreover, dynamic programming can also be used to optimize a grinding process regarding processing time, feeds, and product quality requirements . Machine learning captures critical process parameters with greater accuracy . In laser manufacturing, an Artificial Neural Network (ANN) can be used to predict laser cutting quality, represented by explicit nonlinear functions, to optimize parameters related to laser power, cutting speed, and pulse frequency .
Supply Chain Management.  Recent major disruptions, including natural disasters, geopolitical tensions, and the COVID-19 pandemic, have prompted supply chain managers to seek technologies that enhance sustainability and resilience in order to better address these challenges. In this context, the Supply Chain Digital Twin (SCDT) has emerged as a promising concept, showcasing extensive applications across various sectors. It has the capability to replicate the physical supply chain and identify potential issues before they arise . For instance, the digital twin based on the temperature and quality data can be used to simulate the cooling process in real time. This improves refrigeration processes and reduces food losses, thereby making the refrigerated supply chain greener . Moreover, the port of Rotterdam has been working with International Business Machines Corporation (IBM) to create a digital twin that helps the port test scenarios and understand how to improve operational efficiency . SCDT can also be used to help logistics players manage container fleets more efficiently . AI techniques, including machine learning, evolutionary algorithms, big data analytics, and reinforcement learning, enhance the capabilities of digital twins in supply chains by leveraging historical data to improve real-time data analysis and predictions, thereby expanding the functionalities of SCDT . Anomaly detection is essential for supply chain managers, as it enables the identification of potential issues, allowing for proactive measures to mitigate negative impacts and maintain operational continuity. Machine learning substantially enhances these detection capabilities . Additionally, artificial neural networks can aid in identifying and mitigating risks associated with supply chain operations, such as supply disruptions, demand fluctuations, and shifts in market conditions . Within the SCDT framework, reinforcement learning can optimize real-time decision-making in transportation and logistics, including routing, scheduling, and vehicle dispatch . Furthermore, the automation capabilities inherent in SCDT can be improved through reinforcement learning, empowering stakeholders to make well-informed decisions based on specific scenarios . The quest for optimal inventory policies involves fine-tuning inventory parameters and evaluating inventory costs in relation to service levels. This process can be effectively supported by evolutionary algorithms .
Robotics
Robotics fundamentally revolves around transforming ideas into action, which translates abstract goals into tangible physical outcomes and guiding processes toward their objectives . The use of digital twins in robotics enables a tight integration between physical systems and their virtual counterparts , supporting more precise sensing, modeling, and control. In this regard, digital-twin technologies offer exceptional opportunities for advancing robotics development . Current literature highlights five key domains where digital twins are prominently applied: space robotics, medical and rehabilitation robotics, soft robotics, industrial robotics, and human-robot interaction. Moreover, the integration of artificial intelligence into robotics further empowers systems to become more intelligent, autonomous, and efficient across diverse applications .
Space and Aerial Robotics.  As outer space presents a harsh environment characterized by extreme temperatures, vacuum conditions, radiation, gravity challenges, and vast distances, human access remains both difficult and hazardous. Consequently, human activities in outer space are significantly limited. In this context, space robotics have become essential for assisting these activities . A Virtual Testbed (VTB) specifically designed for optical sensors in space robotics has been developed, which represents a common approach in digital twin applications within space robotics . The researchers in this study aimed to integrate digital twins into the VTB, allowing for the simulation of space robotics while simultaneously controlling the robot actuators in a virtual environment. In practical applications, the digital twin concept has demonstrated its significance by notably enhancing mission efficiency . This improvement stems from its ability to facilitate complex decision-making through the evaluation of simulation outcomes conducted on the digital twin within the VTB . Additionally, a novel simulation-based methodology known as Experimentable Digital Twins (EDT) was proposed . The study also introduced a practical approach for integrating these EDT infrastructures into simulation environments, referred to as virtual test bed (VTB). This integration is believed to hold substantial potential for advancing space robotics, particularly regarding the development and testing of simulation and component algorithms. One study highlighted in presented a haptic telerobotic system that utilized digital twins to seamlessly assemble miniature modular satellites in space. Another research focused on methods for processing the telemetry data associated with the DT-integrated telerobotic system described earlier . The construction, repair, refurbishment, and maintenance of aerospace components, such as spacecraft bodies and systems, have emerged as critical aspects of digital twin adoption in contemporary space robotics. For instance, a robotic grinding system incorporated with digital twin technology was developed for aerospace maintenance, repair, and overhaul. This system employs a 6-DoF robotic arm to perform grinding operations, utilizing the digital twin to analyze and determine essential grinding parameters, such as the required grinding force . Beyond outer space applications, digital twin technologies have also become increasingly essential in the field of aerial robotics and unmanned aerial vehicle (UAV) systems. As UAV swarms grow in scale and operational complexity, digital twins enable high-fidelity airspace modeling, mission rehearsal, and adaptive swarm coordination. Several studies have illustrated these capabilities. For instance, T. Souanef st al. introduced a digital-twin-supported flight safety framework for multi-UAV operations, enabling real-time risk assessment in dynamic airspace . Research presented in  developed a digital twin platform for UAV swarm mission planning, demonstrating improved coordination and reduced collision risks through virtualized testing. In the context of inspection tasks, a digital-twin-enhanced UAV system for infrastructure monitoring was proposed, where the virtual replica was used to evaluate sensor placement strategies and flight trajectories before field operation . These examples demonstrate that integrating digital twins into aerial robotics parallels their use in space robotics, providing a unified simulation-driven framework that enhances autonomy, safety, and mission reliability across both domains.
Medical and Rehabilitation Robotics.  Although the concept of medical robotics is not new, its integration with digital twins has gained significant attention in recent years . One of the fastest-growing areas in this field is DT-aided medical telerobotics, which has key applications in robotic surgeries . Medical telerobotic approaches often utilize immersive virtual reality (VR) interfaces or environments . These include VR-assisted telerobotic medical and laboratory equipment management , telemedical service robots , and RDT-VR-assisted e-skin and soft actuator developments for telerobotic bio-sample collection in contagious environments . Notably, the emergence of the COVID-19 pandemic prompted the development of disaster management-centric telerobotic approaches aimed at preventing infections in such environments . In recent years, another rapidly growing trend in medical robotics has been rehabilitation, which encompasses the development of DT-aided prosthetics, exoskeletons, and other robot-assisted measures. For instance, DT-aided neuromusculoskeletal modeling and simulation can be constructed using CT scans to aid in the development of biomimetic robot prototypes. Additionally, digital twins can simulate and optimize the creation of patient-specific prosthetics , robotic hexapod external fixators focused on correcting bone deformities , and Triboelectric sensor-based exoskeletons . Moreover, a digital twin for a haptic hand exoskeleton can replicate the rehabilitation process in a VR environment . An Automatic Gait Data Control System (AGDCS) has been developed for self-activating DT-aided lower limb exoskeletons . The COVID-19 pandemic accelerated the adoption of robotic assembly for medical equipment to address the escalating demand for such supplies. The digital twin of a robotic system was created for the contactless distribution of medicines and essential supplies in response to the pandemic . Additionally, a DT-optimized Human-Robot Collaboration (HRC) system was designed to meet the increasing needs for medical equipment, particularly alternators and ventilators . The implementation of AI in conjunction with digital twins has also been on a steady rise. For example, R-CNN has been employed to detect various medical equipment and their corresponding digital twinsâ positions within a virtual environment . That same year, a DT-integrated robot platform was developed utilizing deep learning to automatically collect bio-samples from patientsâ nasal vestibules . Lastly, micro-nano medical robotics represents an emerging trend in the field. In 2022, a digital twin of a micro-bot was developed, which employed AI to predict system outputs.
Soft Robotics.  As we know, sensors are employed to monitor both their movements and external stimuli, particularly in fields such as surgical and micro-nano manipulation, where accurate motion detection and tactile sensing are crucial . Soft robotics, characterized by high compliance and dexterity with muscle-like actuators made from materials such as silicone rubber and thermoplastic polyurethanes (TPUs) , is well-suited for collaboration with sensors to enhance monitoring and control in robotic applications. Although soft robotics is a relatively new field, it holds considerable potential. The digital twin concept in soft robotics generates digital information that accelerates development . Recent research trends in the digital twin realm of soft robotics are increasingly focused on augmented and extended reality . Soft robots can serve as virtual humans within human-centered production systems, where infrastructure efficiency can be further enhanced by leveraging digital twins in virtual reality (VR) environments . A digital twin of a soft robot utilizing pneumatic muscles has been developed, offering numerous applications, including cellular production, where it optimizes the work environment and reduces space requirements . Traditional data processing techniques are insufficient for managing the heterogeneous big data necessary to establish human-machine interfaces (MMI) for the digital twin. This necessitates the use of machine learning approaches in conjunction with advanced communication protocols like 5G to maximize the efficiency of digital twin-soft robotics and related systems . For example, machine learning enhances data interpretation for improved manipulation or detection, such as accurate gesture recognition . This approach is used to explore the capabilities of specially designed triboelectric nanogenerator (TENG) sensors. A tri-actuator soft gripper, fabricated through three-dimensional (3D) printing and integrated with TENG sensors, perceives gripping status and identifies objects by leveraging machine learning for data analysis. Additionally, a digital twin framework is established to create a duplicate digital representation of the aforementioned manipulation within a virtual reality (VR) environment, often referred to as cyberspace .
Industrial Robotics.  The industrial robotics sector is one of the fastest-growing divisions, offering standardized technologies suitable for various automation processes . A comprehensive analysis in revealed that the digital twin could be effectively utilized for five types of industrial plant maintenance: reactive, prescriptive, condition-based, predictive, and preventive maintenance, with the latter two being particularly effective. These approaches are also applicable to robot-integrated systems. Vachålek et al. demonstrated a digital twin that employed a genetic algorithm for DT-optimized predictive analysis and plant augmentation in a robot-assisted production line . Subsequently, proposed a Robot Digital Twin (RDT) modeling methodology utilizing a physics-based approach and virtual sensors to collect and generate data from industrial robots for predictive analysis. In 2018, a reinforcement learning approach was applied to enable a robot to learn to lift various weights autonomously. The robot model could be visualized and controlled through its digital twin . More recent developments include DT-integrated machine vision assessments of industrial robotic skills and enhanced industrial robot programming using machine learning with point cloud information and RDT . A significant limitation of the data-driven RDT approach, as highlighted by , is the lack of historical data. However, this challenge can be addressed by generating synthetic data, which can be further improved using AI algorithms. In , the authors presented a deep learning approach for generating additional synthetic image data from digital models, achieving a 100% success rate when the developed model was retrained and utilized for detecting real-world objects in various orientations. Developing efficient AI algorithms for physical robots poses a critical challenge due to the excessive time consumption, power supply, and component constraints for long-term repetitive tasks, as well as the absence of suitable virtual testbeds (VTBs). An effective solution lies in developing algorithms through extensive simulations, where digital twins have proven to play a vital role . Notable examples include DT-aided deep reinforcement learning policy transfer from simulation to physical robots , and training robots within their digital twin for intelligent grasping using a grasp-generation-and-selection convolutional neural network, which achieved 96.7% and 93.8% success rates for gripping single items and mixed objects, respectively .
Field and Service Robotics.  Field and service robotics span diverse application domainsâincluding agriculture, construction, mining, underwater exploration, search-and-rescue, logistics, and public serviceâwhere robots must operate in dynamic, unstructured, and often hazardous environments. Such settings introduce substantial uncertainties arising from terrain variability, environmental disturbances, adverse weather, sensor noise, and unpredictable human interactions. Digital twins have emerged as a powerful tool to mitigate these challenges by providing high-fidelity virtual counterparts of physical systems for analysis, simulation, planning, and control . In agriculture, Melesse et al. reviewed DT-assisted systems that model crop growth, soil moisture evolution, and canopy structures to support autonomous weeding, fruit picking, and targeted spraying . L. Yining et al. further introduced DT-aided robotic harvesters equipped with multispectral sensing, in which the digital twin supports fruit-state estimation and enables adaptive manipulation of delicate crops . In construction robotics, DT-enabled frameworks integrating building information modeling (BIM), real-time site scanning, and autonomous machinery control have been demonstrated . DT-based models have been applied to generate optimized excavation trajectories for autonomous earthmoving systems, while DT-supported robotic construction workflows improve placement accuracy through continuous synchronization between virtual and physical models . Additional efforts include DT-driven autonomous rebar tying , concrete finishing robots developed and validated in DT-based simulation environments , and site-inspection robots exploiting virtual replicas for collision-aware navigation in cluttered and evolving workspaces . In mining and underground environments, DT technologies support higher-risk operations. Lee et al. introduced digital twins of subterranean tunnels that integrate geological models, environmental sensing, and operational records to enhance situational analysis and planning . Complementary studies show that DT-assisted systems can simulate rock deformation processes or ventilation dynamics prior to execution, enabling improved safety assessment and operational efficiency . Underwater robotics imposes additional constraints, including limited visibility, dynamic water currents, and severe communication restrictions . Orjales et al. developed DT-based underwater simulation environments with high-fidelity hydrodynamic modeling, allowing autonomous underwater vehicles (AUVs) to refine navigation and manipulation strategies under realistic virtual conditions .
Human-Robot Interaction.  The study of Human-Robot Interaction (HRI) has been a focal point of extensive scientific research over the past decades . The enhanced capabilities provided by digital twins significantly expand the potential for leveraging human-robot systems effectively. Consequently, digital twins in HRI are drawing significant attention for future innovations . Early efforts in this field aimed to develop reliable and interactive robot control interfaces and affordable 3D reconstructions of digital twins for robot control in factories . Digital twin-aided robots in virtual reality (VR) and mixed reality (MR) have emerged as fast-growing research trends in recent years. Applications include HRI workspace design and optimization using VR , robotic construction supervision in MR , and robot programming in VR and MR . Other notable applications of digital twin-aided robots include warehouse and indoor automation using Autonomous Mobile Robots (AMRs) , and digital twin-integrated, energy-efficient smart manufacturing . Gesture control methods have also been notably applied in various telerobotic applications . The Leap Motion method, due to its easy implementation in VR and digital twin integration, has become a popular choice for hand gesture-controlled robotic systems . Robots are made aware of collisions with workplace objects and human operators through real-time simulation and processing in a virtual test bed (VTB), where digital twins represent the robots, human operators, and workspace . Many approaches utilize the versatility of MR/VR interfaces to further augment digital twin-aided telerobotics . These augmentations include establishing immersive user interfaces for robot control , operations in hazardous environments like nuclear facilities , kinesthetic aid to robot operators , and AI-based VR data processing for improved robot control . Clone digital twins, which collect extensive data from their physical counterparts, represent a trendy application in human-robot systems. They collect near-real-time sensorial data from humans, robots, and their co-environment . The scalable data exchange characteristic of clone digital twins allows for incorporating intelligent features using AI and ML for planning, optimizing, and automating human-robot interactions . Furthermore, emerging haptics technology provides an additional data source for HRI . This technology, when combined with Extended Reality (XR), enables real-time sensory experiences for humans, significantly improving joint human-robot task execution. Recent studies on digital twin-AI applications in human-robot interaction (HRI) have focused on various areas, including machine learning and deep learning techniques for robot trajectory estimation and obstacle detection, which ensure safe and collision-free HRI workspaces . Additionally, Fast Fourier Transform (FFT) and machine learning-based digital twins are utilized in human-robot interactive welding and welder behavior analysis . Reinforcement learning-based autonomy in complex assembly environments aims to reduce operator fatigue . Another notable development involves lifecycle frameworks and the optimization of pick-and-place robots for virtual product development . Furthermore, DT-aided deep learning is being applied for human action recognition , alongside data augmentation techniques for VR-assisted tele-manipulation and eye-gaze and head gesture recognition systems that facilitate gesture control in robot tele-manipulation . For instance, Mobile ALOHA features a low-cost full-body teleoperation system that collects data from high-quality human demonstrations, alongside a novel imitation learning algorithm that effectively learns from these demonstrations . Finally, significant advancements have been made in robots capable of self-learning assembly processes .
Natural System and Environment
The natural environment is crucial for sustaining life on Earth and maintaining ecological balance, exerting a profound influence on human activity . Environmental concerns have sparked numerous important topics of discussion, including climate change , biodiversity , conservation efforts , and sustainable development . Digital twin models, benefiting from their unique ability to create virtual replicas of environmental entities and simulate cyber-physical systems, are gradually gaining attention as a promising solution to these environmental challenges . While the digital twin technology is well-established in industrial sectors, its applications in environmental sciences remain nascent due to the complexity of modeling environmental systems . Traditional physics-based dynamical models from earlier eras are complex and computationally expensive, failing to support high-resolution, interactive digital twin systems for environmental simulations . However, by adopting machine learning and artificial intelligence models, digital twin systems are evolving to offer greater scalability, interoperability, and high-fidelity environmental cyber representations . This integration not only facilitates rapid interactions among physical environments but also accelerates simulation processes, promoting academic and industrial innovations in fields such as weather , ocean , geology , wildlife , and forestry . A frontier view of digital twins proposed by researchers, organizations, and companies envisions the creation of a digital Earth, which is portrayed as a dynamic, interactive replica of Earthâs past, present, and future states .
Weather Prediction.  Weather prediction aims to reduce associated losses and improve societal benefits by providing accurate information on weather conditions . Digital twin models offer precise, high-resolution predictions by combining machine learning and deep learning techniques , making them pivotal in various applications. Research has demonstrated that incorporating digital twin models in weather prediction can efficiently support decision-making processes in natural disaster risk mitigation , weather services , and renewable energy grids . To predict natural climate disasters, researchers have developed numerous digital models. Brocca envisioned the Hydrology model, a digital twin of Earthâs hydrological processes, to predict flooding by utilizing real-time data from both terrestrial and satellite measurements of rainfall and soil moisture, aiding in disaster mitigation . Similarly, a digital twin model based on both unsupervised and supervised learning was constructed to monitor ice storms dynamically. Efforts to improve weather services through digital twin models are also underway. Singh et al. proposed a framework for generating high-resolution urban precipitation data to revolutionize urban climate services, providing city planners with timely and precise climatic information . Koldunov and Jung demonstrated how LLMs can deliver localized climate services by combining the strength of LLMs and model simulations, making manipulations in simulations more accessible . Chen et al. incorporated vision information into LLMs, proposing a novel Vision-Language Model approach . This integration considerably enhances the analysis of weather heatmaps, improving the speed and accuracy of extreme weather event detection based on the predictive capability of LLMs. Furthermore, weather conditions play a crucial role in energy production. Sehrawat et al. developed a digital twin system for predicting solar irradiance , while Stadtmann et al. designed a digital twin aimed at enhancing wind energy production efficiency . Savage et al. focused their digital twin research on facilitating collaboration between energy and climate research .
Ocean System.  The ocean system is essential to life on Earth, making the understanding of ocean dynamics profoundly significant . Digital twin models for the ocean cover diverse areas, showcasing their vast potential in understanding the state and changes of the ocean, such as ocean sustainability , ocean observation , and coastal flood predictions . Digital twins play a crucial role in ocean sustainability by reducing and preventing overfishing, modeling and predicting marine pollution and facilitating marine spatial planning . Rakotonirina et al. conducted a digital twin study for Ocean Cleanup systems, optimizing system design, predicting fleet performance, and estimating operational costs . In the application of observing and simulating the ocean, an underwater network of ocean observation systems has been deployed in the Baltic Sea, demonstrating the feasibility of digital twins in extreme underwater environments . Similarly, Zheng et al. enhanced wave height predictions for ocean energy generation by integrating recurrent neural networks . OceanNet, a physics-informed neural network-based digital twin, models and predicts ocean circulation in regional seas using a Fourier neural operator, promoting stability and mitigating autoregressive error growth over long-term forecasts . Coastal flood prediction is another notable application. A local digital twin earth project on the Italian coast integrates environmental, demographic, and marine data into a high-precision digital model to monitor, simulate, and predict natural and human activities . Thiria et al. developed an advanced deep learning method for enhancing the spatial resolution of ocean current estimates from satellite observations by merging low-resolution geophysical ocean fields with high-resolution data using convolutional neural networks . Physics-informed machine learning techniques have also been used to create fast and accurate models for coastal flood predictions, significantly accelerating the simulation process while maintaining high accuracy in predicting sea surface height .
Geological Simulation.  Simulation in geology is a vital and long-lasting topic that has gained prominence due to the inherent complexities of underground processes, which are often unobservable directly . Digital twin models have transformed traditional geological simulations due to their precision and efficiency , stimulating the development of areas such as geological modeling , underground operations , and geological disaster monitoring . In the aspect of geological modeling, Wang et al. employed a digital twin framework to construct complex 3D geological models, incorporating real-time data to refine these models and adapt to new geological conditions . Yang et al. introduced the EdGeo toolkit, a physics-guided generative AI tool for subsurface exploration, which enhances the fidelity of velocity maps through diffusion-based models . Besides, digital twin models substantially aid underground operations. In tunnel construction, dynamically updatable digital twin models are used to manage the continuously evolving geological conditions encountered in long-term projects . In the mining sector, the digital twin system provides updated and precise representations of deposits, supporting informed decision-making in mining operations . Additionally, digital twin models are instrumental in monitoring geological disasters. Zhang et al. utilized digital twins for real-time, dynamic predictions of complex and frequent geological disasters, such as landslides, earthquakes, and mudslides, in their monitoring and early warning systems . For seismological research, Vladimir et al. implemented numerical methods and high-performance computing in a digital twin system to create detailed 3D models of geological structures, simulating seismic wave propagation to improve subsurface understanding .
Wildlife Protection.  Biodiversity possesses inherent value for the natural world and is crucial for maintaining ecosystem functionality . Wildlife plays a vital role in supporting biodiversity . Digital twins can simulate wildlife and provide dynamic, predictive insights that enhance understanding of ecological systems . These insights include monitoring animal behaviors and population status, thereby strengthening conservation efforts. Sharef et al. developed an interactive machine learning framework integrated with digital twins to improve biodiversity projection models . A prototype simulating virus spread among wildlife, aiding in effective management strategies, has been created by Ingenloff . To mitigate wildlife-vehicle collisions, Moulherat et al. designed an integrated system using sensors and machine learning, leveraging real-time data from camera traps for ongoing wildlife management . Fergus introduced an innovative system where digital transactions triggered by camera trap detections compensate local guardians for their conservation efforts, fostering community involvement in biodiversity preservation . Rolph expanded the application of digital twins to the management of cultural ecosystem services, merging biodiversity and community recreation to provide a comprehensive assessment of ecosystem services . Sakhri adapted digital twin technology to optimize energy use in monitoring waterbirds, demonstrating significant improvements in real-time monitoring capabilities . Teschner introduced a novel digital twin-enhanced approach using unmanned aerial vehicles to protect agricultural fields from wildlife intrusion, showcasing the technologyâs adaptability to various conservation needs .
Forest Management.  Forests contain necessary functionality in maintaining the health of Earthâs ecosystems and climate . Digital twin models have the ability to utilize historical data to simulate and forecast changes in forest ecosystems over time . Hence, the advent of digital twin technology has brought transformative changes to the field of forest management, marking a significant advancement in how forest ecosystems are monitored, analyzed, and managed . Dynamic ecosystem modeling by Qiu et al. utilizes remote sensing and 3D parametric modeling for real-time forest ecosystem monitoring . At the same time, Li et al. have developed a robust database for virtual plantation management . These technologies enable a highly interactive and dynamic digital twin model for forests that allows for real-time monitoring and decision-making, enhancing the management of forest ecosystems by enabling synchronization between virtual and real interactions. On a national scale, Li et al. utilized a multi-task deep learning network to accurately map and characterize every single tree within and outside forests across Denmark, identifying the location, crown area, and height of trees, thereby enhancing national forest management and conservation efforts . Expanding further, MĂ”ttus et al. introduced a high-precision global model under the Digital Twin Earth initiative, enhancing vegetation mapping and characterization across Europe . Urban forests also benefit from digital twin technologies, as demonstrated by Ozel and Petrovic. Their Green Urban Scenarios framework incorporates factors such as weather conditions, tree species, diseases, and spatial distributions into simulations, helping to predict the future trajectories and impacts of urban forests under different scenarios . Innovations continue with the introduction of adaptive management strategies using machine learning. Damavsevivcius have applied reinforcement learning algorithms within their digital twin models to optimize forest management . Additionally, Zhong and Sanchez-Guzman both focus on wildfire management, utilizing digital twin models to predict and manage wildfire dynamics effectively, providing real-time data that helps prevent and mitigate forest fires.
Digital Earth.  While researchers have developed digital systems and machine learning models at national, regional, and local levels , the arrival of more advanced deep learning models and powerful computational technology has made the vision of a global digital twin of Earth increasingly feasible. This ambitious project aims to realize global climate projections and assess local impacts simultaneously. It integrates advanced computer science, mathematics, and engineering with interdisciplinary knowledge from natural sciences to create global-scale interactive models of Earthâs systems . This vision has sparked widespread global discussion . The Earth-2 project exemplifies this ambition by simulating real-time, global-scale meteorological conditions and constructing a replica of the entire Earthâs atmospheric system to predict weather conditions, natural disasters, and conduct meteorological explorations. Data-driven physics-informed machine learning and deep learning models strive to embed the rules of the physical world into their frameworks . A notable success in this field is the FourCastNet model, which employs Adaptive Fourier Neural Operators to revolutionize global weather forecasting with high-resolution outputs . Building on this, Kurth et al. have demonstrated the scalability and efficiency of FourCastNet on supercomputing systems, paving the way for large-scale, real-time, high-resolution global weather forecasting .
Agriculture
As the global population surges and climate change impacts food security, there is a pressing need for agricultural systems to increase production efficiency while minimizing resource consumption. In response, digital twin technology has become increasingly essential due to its advantages. Digital twins, which are synchronized virtual counterparts of physical objects or systems, offer significant potential to transform agricultural practices by providing detailed virtual replicas of farms and agricultural objects . These virtual agriculture systems promote smart farming, livestock management, and agricultural facility optimization, thereby offering advantages such as cost savings, improved product quality, and enhanced operational efficiency to effectively address agricultural challenges . Machine learning and deep learning technologies are being implemented across all facets of agriculture, spanning preproduction, mid-production, and postproduction phases, enabling more accurate simulations . Researchers highlight that artificial intelligence facilitates data processing and analysis in agricultural digital twin systems, supporting decision-making and providing feedback to virtual systems . Contemporary advancements in generative AI have the potential to augment digital twin technologies, addressing longstanding challenges such as farmer-system interactions and agricultural data synthesis . The integration of artificial intelligence with digital twin technology is propelling significant advancements in agriculture, enhancing the potential for increased crop yields while simultaneously reducing environmental impact and optimizing resource utilization .
Smart Farming.  Smart farming, a paradigm shift in agricultural practices, leverages an array of modern innovations, including cloud computing, the Internet of Things, machine learning, augmented reality, and robotics to revolutionize agricultural production . As farms increasingly embrace digitalization, the concept of digital twins has emerged as a comprehensive framework that virtualizes every object within the farm, creating digital replicas of crops that mirror their behavior and states throughout their lifecycle in a virtual space, enabling farmers to optimize operations . In the realm of irrigation management, Alves et al. have pioneered a sophisticated digital twin system that integrates soil, weather, and crop data to generate daily irrigation prescriptions . Nitrogen application, another critical aspect of farming, has also benefited from digital twin technology. Stefano et al. discovered that employing digital twins to simulate diverse scenarios and analyze data empowers farmers to make more informed decisions regarding nitrogen application . Smith emphasizes the applications of AI in enhancing precision when detecting and measuring farm activities to enable reliable alerts for farmers . Visual neural network techniques have proven particularly effective in monitoring plant health. Nasirahmadi et al. demonstrated the power of coupling these techniques with other smart farming technologies to detect and assess plant health with remarkable accuracy . Integrating multiple smart farming technologies, Angin et al. proposed AgriLoRa, an innovative digital twin framework, which utilizes cloud-based computer vision algorithms to identify plant diseases and nutrient deficiencies, providing farmers with actionable insights to improve their crop management strategies .
Livestock Management.  The field of livestock management is another important area in agriculture, facing increasing challenges due to finite resources, the need to reduce greenhouse gas emissions, and a declining global workforce . Digital twin technology has emerged as a promising solution in livestock farming. This technology creates a digital replica that simulates the physical, biological, and behavioral states of animals based on real-time data input, enabling optimal and sustainable livestock management operations . The integration of digital twins with augmented reality (AR) and virtual reality (VR) technologies further enhances the training of veterinary professionals and breeders by providing immersive, interactive learning environments that simulate real-world scenarios, ultimately improving animal welfare and operational efficiency . In the realm of aquaculture, AI-powered IoT digital systems are revolutionizing fish farming practices by employing sensors and smart devices to collect real-time data on fish metrics, environmental conditions, and health status, enabling automated fish feeding, water quality monitoring, and disease detection . Advanced research in this area includes replicating the hydrodynamic behaviors of living organisms in aquatic environments using visual neural networks that extract features from video data, combined with computational fluid dynamics to analyze and predict the hydrodynamic cues for fish navigation . Urban beekeeping has also benefited from digital twin technology. A multiagent model integrating bee populations, beekeepers, non-beekeepers, and the environment has been developed to monitor and analyze bee colony health and behavior, providing a robust tool for decision support for small-scale farmers and urban planners . For cattle management, an innovative digital twin model powered by LSTM neural networks has been developed within a farm IoT system to monitor and track cattleâs physiological and behavioral states in real time, leveraging the power of artificial intelligence to predict future behaviors and physiological cycles of cattle .
Controlled Environment Agriculture.  Controlled Environment Agriculture (CEA) stands in stark contrast to traditional farming methods, focusing instead on artificially constructed agricultural facilities. These facilities, which include physical structures and installations, are designed to foster optimal conditions for enhanced farming quality and efficiency, thereby optimizing resource utilization . Serving as the cornerstone of modern large-scale agriculture, these systems employ advanced monitoring and control technologies . They show immense potential when integrated with Digital Twin (DT) technology, promising to invigorate agricultural practices . Digital twin systems excel in supervising conditions within enclosed structures such as greenhouses or indoor facilities. These sophisticated systems manage and manipulate various environmental factors, including temperature, humidity, light intensity, CO2 levels, and air flow . By constructing digital replicas of physical farms, a digital twin system can simulate and optimize environmental conditions, enhancing animal welfare and improving farm management . Further advancements in CEA are geared towards increasing energy efficiency and operational effectiveness, particularly in livestock management. A proposed digital twin framework, for instance, replicates physical pig houses in a virtual environment. This enables simulations that optimize heating, ventilation, and air conditioning systems . In commercial greenhouse production, digital twin technology optimizes energy usage and streamlines production processes. Howard et al. have described a digital twin system that integrates climate control, energy management, and production processes to simulate and refine operational strategies without disrupting ongoing cultivation . Prawiranto et al. conducted a comprehensive study on optimizing solar drying processes for fruits. Their approach utilizes a physics-based Digital Twin that incorporates mechanistic fruit drying models, quality models, and weather data to evaluate different improvement strategies .
Commerce
Traditional live commerce, while rapidly growing, faces numerous challenges such as crowding, limited operating hours, and long queues. It also suffers from several limitations, including unengaging content and limited interactivity . Digital twinning enables the creation of virtual replicas of physical objects, allowing consumers to interact with products or services in a virtual space that closely mimics real-world conditions . This innovative technique can transform both e-commerce and retail stores by constructing digital twins of products and customers, thus enhancing the shopping experience . Additionally, its capacity to facilitate collaborative design permits multiple stakeholders such as customers, designers, and manufacturers to engage in the product development process through the convergence of physical and virtual data . Furthermore, with a digital representation of financial data and activities integrated with the capabilities of Large Language Models, a digital twin system can also provide financial services to individuals, institutions, and governments . Unlike traditional commerce, digital twins utilize real-time data and employ machine learning models . Deep learning algorithms, such as image recognition, 3D reconstruction, and even generative AI, are used to categorize products, optimize shop layouts, and predict customer needs more accurately . These technologies enhance operational efficiencies in customization and responsiveness, promoting informed decision-making and improving customer experiences in both virtual and physical stores.
E-commerce.  In the evolving landscape of e-commerce, the integration of digital twin technology has introduced novel applications ranging from immersive shopping experiences to advanced fraud detection and realistic simulations of consumer behavior . Jeong and colleagues have pioneered the MBUS platform, combining metaverse elements with live commerce to create a virtual shopping environment, enhancing user engagement by allowing real-time interaction with products . Simultaneously, Wang et al. have advanced fraud detection by modeling user behaviors as genetic sequences in the SAGE framework, which employs principles from genetics to enhance the detection of fraudulent activities . Yao and his team have contributed to WebShop, a large-scale interactive environment that trains language agents to perform web-based tasks by navigating and customizing real-world product interactions based on textual instructions . Terån and his collaborators have developed an agent-based simulation model that incorporates word-of-mouth dynamics and endorsement theory to more accurately represent market behaviors . Furthermore, Kuzmichev explores the application of digital twins in the fashion industry, focusing on 3D digital garment design . This process significantly reduces labor and material costs by using virtual human models and virtual garments to simulate the interaction between the body and clothing. Fu et al. developed an innovative Augmented Reality try-on system for virtual clothing digital twins, featuring real-time interactions and realistic cloth simulation through an optimized framework, which has the potential to significantly enhance customer experiences in virtual fashion retail environments .
Retail.  The collection and analysis of real-time data has become the engine for continued growth in the retail industry . Digital twin technology can serve as a framework for data integration, providing intelligent analysis for various aspects of the retail industry and facilitating decision-making . Vijayakumar explores the concept of a behavioral digital twin, which models consumer behaviors to tailor interactions and predict future purchases, aiming to significantly enhance customer satisfaction through personalized experiences . In a different application, Shoji et al. focus on the postharvest life of imported fruits, utilizing physics-based digital twins to control and monitor hygrothermal conditions, thus improving fruit quality from packhouse to retail stores . Sengupta and Dreyer demonstrate how digital twins can facilitate sales and operations planning in grocery retail, helping to predict and manage variabilities across the value chain to minimize waste . Additionally, Liu et al. present a multi-modal approach that employs AI-driven real-time product recognition and 3D store reconstruction to create a precise virtual representation of physical retail spaces, enhancing both operational efficiency and customer interaction . Pous et al. introduce an innovative use of robots that scans item locations within a store, integrated into a digital twin, providing customers with real-time product information . Stacchio et al. discuss the social acceptance and potential benefits of Human Digital Twins in fashion retail, highlighting how 3D models of customers can improve service and customer satisfaction in brick-and-mortar stores .
Financial Service and Analysis.  The integration of digital twin technology in financial management and economic simulations is reshaping the capabilities of existing tools by incorporating real-time data, LLMs, and more human-like decision-making processes. Anshari and colleagues discuss the application of DT technology within robo-advisors, transforming them from static tools into dynamic financial advisory platforms. This integration allows for continuous improvement through interaction and real-time data updates, providing users with more tailored and responsive financial management solutions . Li et al. present EconAgent, an agent-based modeling framework that leverages LLMs to simulate macroeconomic phenomena, enabling more realistic decision-making based on work, consumption, and past economic trends . In another innovative approach, the Agent-based Simulated Financial Market utilizes LLMs to mimic real human traders and simulate stock market behavior. The ASFM framework includes a realistic order matching system and simulates various industry sectors, creating agents with diverse profiles and strategies that can understand market dynamics and respond to economic news . Veshneva formulated a novel approach for constructing digital twins of socio-economic systems using status function-based mathematical models, enhancing predictive analytics and decision-making in complex, uncertain environments . Barkalov et al. envisioned an innovative distributed forecasting information system architecture integrating digital twins and recurrent neural networks for predictive asset maintenance in socio-economic systems, demonstrating improved accuracy in predicting remaining useful life and potential asset failures .
Education and Training
Over the past four years, the digital twin technique in education has gained popularity, driven by safety concerns associated with traditional offline education during the COVID-19 pandemic . As virtual modules that replicate real-world construction processes, digital twins can blend online and offline learning, emerging as a solution to ensure the continuity and quality of education . Some researchers argue that due to the rapid acceleration of knowledge doubling, traditional education methods are becoming limited . The dynamic digital twin model facilitates an integrated educational environment, merging physical and digital spaces to provide comprehensive educational resources, such as traditional knowledge, experience, and societal wisdom. It also enhances tailored learning experiences by integrating various aspects of a studentâs academic and personal life . The digital twin system enhances learning outcomes by enabling learners to experience efficient data communication and interaction, facilitating the understanding of complex, practical knowledge through virtual reality integration, as demonstrated by existing results . Furthermore, by leveraging advanced technologies such as the Internet of Things, artificial intelligence, virtual reality, and 5G, educational institutions can create more effective and personalized learning environments . Recently, more articles have demonstrated the potential of generative AI, especially LLMs and agents, as revolutionary tools to boost simulations and enhance educational methods .
Virtual Learning.  Applications of digital twins in educational settings, ranging from immersive virtual environments to hands-on learning tools, significantly enhance student engagement and understanding across various disciplines . For instance, Nikolaev et al. implemented a digital twin system that enables MSc students to create virtual models of tunnel boring machines, enriching the learning experience and sparking increased interest in the program . Madni et al. have engaged students in developing digital twins by instrumenting physical vehicles with sensors, helping students understand vehicle dynamics and control techniques, and moving beyond traditional lecture-based learning . Furthermore, Han et al. have developed a comprehensive digital twin of campus using UAV tilt photography and 3D modeling, integrated via Unity3D to facilitate real-time monitoring and decision-making . Razzaq et al. have pioneered DeepClassRooms, a digital twin framework that utilizes convolutional neural networks for enhanced attendance tracking and content delivery monitoring in public sector schools, demonstrating the adaptability of digital twins in addressing educational challenges in resource-limited settings . Zhang and colleagues have introduced SimClass, a cutting-edge framework that simulates traditional classroom settings using LLMs. This framework improves educational experiences by allowing LLMs to assume various classroom roles, which promotes cognitive and social development as well as collaborative behaviors, enriching overall classroom dynamics .
Interactive Learning.  Digital twin technology in education is advancing through innovative applications to foster collaborative and interactive learning environments . Digital twin systems use data analytics and AI to adjust the content difficulty, pacing, and learning paths in real-time, personalizing the learning experience based on individual student performance and needs . Sepasgozar has employed these technologies to allow students to safely explore complex construction activities through the VTBM module, an immersive virtual environment . Innovatively, Lee and colleagues have introduced game-like elements within the digital twin framework to demystify mathematical concepts and engage students in problem-solving activities . In the holographic classroom envisioned by Liu and Ba, digital twins provide an immersive and interactive 3D learning space, merging physical and virtual worlds to enhance educational experiences and introduce dynamic teaching methods . Lan and Chen have developed a team-teaching framework where human educators are paired with AI agents, designed to support personalized learning and real-time feedback. This approach enhances pedagogical effectiveness while preserving the dynamics of human teaching . In mathematics education, Yue et al. have developed MATHVC, a virtual platform where LLMs simulate student interactions in a multi-character setup. This platform encourages collaborative problem-solving and offers a scalable method to practice mathematics without continuous teacher supervision, thereby enhancing student engagement and learning autonomy . Murtaza and his team explore the application of ChatGPT in driver education, comparing traditional instructional methods with interactive, ChatGPT-based learning. Their findings indicate that participants trained with ChatGPT exhibit significantly better learning outcomes, demonstrating the effectiveness of LLMs in practical applications .
Quantum Computing for Digital Twin
Motivation: Why Digital Twins Need Quantum Computing.  As digital twin (DT) technologies evolve toward modelling ever more ambitious systemsâmulti-physics, high-dimensional, real-time and tightly coupledâthe computational burden of simulation, inference, optimisation and control escalates rapidly. Many core DT workloads can be cast as high-dimensional linear algebra, stochastic simulation and large-scale combinatorial optimisation problems, whose classical complexity typically grows polynomially or even exponentially with the relevant state, parameter or decision-space dimension. Even with modern high-performance computing (HPC) and accelerator hardware, these scaling limits constrain both the achievable spatial/temporal resolution and the breadth of scenario exploration within operational time budgets .
Classical digital twin architecturesâtypically comprising sensor-data ingestion, AI/ML-based surrogate modelling and numerical simulationâtherefore begin to hit intrinsic bottlenecks in three key areas, even when backed by large-scale HPC infrastructure:
1. Scale and Coupling: Systems such as whole smart-cities, biological networks, aerospace systems or large manufacturing plants involve high-dimensional state-spaces, strong non-linearities, multi-domain physics (mechanical, thermal, electrical, biological), and real-time feedback loops. In practice, this translates into repeatedly solving large coupled PDEs/DAEs, high-dimensional Bayesian inference problems, and global or mixed-integer optimisation problems over many control and design variables. The complexity of modelling, simulating and controlling such systems at high fidelity is daunting for purely classical methods and leads to severe trade-offs between resolution, coverage and latency.
2. Real-time/Near-real-time Operation: Digital twins are increasingly used not only for offline analysis but for operational decision-making, fault detection/prediction, autonomous control and adaptation. Achieving low-latency, high-fidelity simulation and inference across large systems is challenging for classical methods, especially when decisions must be updated at time scales comparable to measurement rates. In the near term, any quantum acceleration is more realistically applied to computationally intensive but latency-tolerant subroutines (e.g., periodic re-optimisation, scenario generation, policy improvement), which indirectly support real-time DT operation rather than sitting directly in millisecond-level feedback loops.
3. Complex Optimisation and Inference Under Uncertainty: Typical DT tasks include combinatorial optimisation (scheduling, routing, control), probabilistic inference with uncertainty quantification, inverse problems (e.g., deducing system state or faults from sensor data), and high-fidelity simulation of physical phenomena. These tasks map naturally onto classes of quantum algorithms that promise asymptotic speedups in query or sample complexity under suitable assumptions: quantum approximate optimisation algorithms (QAOA) and related variational methods for combinatorial problems, quantum amplitude estimation and quantum-accelerated Monte Carlo for uncertainty quantification, and quantum linear solvers and PDE solvers for large-scale simulation and inverse problems . However, rigorous end-to-end quantum advantages for full DT workflows remain an open research question, particularly on noisy intermediate-scale quantum (NISQ) hardware.
In this landscape, quantum computing (QC) is not a monolithic replacement for classical DT infrastructure but a prospective source of algorithmic acceleration for specific computational kernels. Rather than focusing on physical phenomena such as superposition and entanglement, it is more useful at the DT level to view QC through the lens of quantum algorithms for simulation, optimisation and learning, many of which admit provable or conjectured complexity-theoretic advantages over the best-known classical counterparts . We use the term Quantum Digital Twin (QDT) to denote a hybrid quantumâclassical DT architecture in which classical components manage sensor interfaces, data pre-processing, domain-specific logic and most control tasks, while quantum processing units (QPUs) are invoked as accelerators for carefully selected âhard-coreâ subproblems (e.g., large-scale optimisation, high-dimensional inference, or fine-grained physical simulation) . Recent position papers and industrial case-studies have started to articulate such QDT architectures and early prototypes in domains including manufacturing, smart-city operation and supply chains .
Beyond simulation and optimisation, QDTs may incorporate quantum machine learning (QML) or quantum AI (QAI) modules as high-capacity function approximators for perception (feature extraction from high-dimensional sensor data), prediction (state and parameter forecasting) and decision-making (policy learning and planning). QML approaches based on variational quantum circuits and quantum feature maps are being actively explored as a way to trade classical sample or computational complexity for quantum circuit depth and qubit resources . At the same time, NISQ-era devices impose strict limitations in terms of qubit count, noise and connectivity, so near-term QDTs are likely to deploy quantum modules in offline or batch modes and to rely heavily on hybrid schemes such as variational quantum algorithms (VQAs), which combine parameterised quantum circuits with classical optimisation .
From an architectural perspective, it is useful to distinguish several generic types of quantum modules that can be reused across DT domains: (i) quantum simulation modules for high-dimensional linear systems, PDEs or quantum chemistry; (ii) quantum optimisation modules for combinatorial and continuous optimisation; (iii) QML/QAI modules for supervised, unsupervised and reinforcement learning; and (iv) quantum-enhanced cryptography and secure communication primitives. Each application domain below can then be viewed as instantiating different combinations of these generic modules according to its dominant computational bottlenecks.
Quantum Digital Twins Across Application Domains.  Below we follow the domain structure of Sections 7.1â7.11 and discuss how quantum computing can enhance digital twins in each domain. For each, we highlight the key potential and map DT workloads to candidate quantum modules, citing relevant literature where available. In several domains, these mappings remain largely conceptual, reflecting the early stage of QDT research and deployment.
1. Healthcare Systems
Digital twins in healthcare simulate patient-specific organs, physiological systems, treatment planning and longitudinal monitoring. When augmented with quantum computing, at least three main enhancements can be envisaged. First, in the longer term, quantum chemistry and quantum molecular simulation algorithms could improve the fidelity of drugâtarget interaction models and thereby supply more accurate mechanistic priors or parameter sets to patient-level DTs for personalised medicine . In the near term, such quantum simulations are more likely to be used offline for molecular screening and model calibration rather than in real-time clinical workflows. Second, QML/QAI models may be used to analyse heterogeneous, high-dimensional patient dataâsuch as longitudinal electronic health records, imaging and multi-omicsâto predict disease trajectories and treatment responses under uncertainty, for example via quantum kernel methods or variational quantum classifiers . Third, treatment scheduling, radiotherapy planning and hospital resource allocation can be formulated as large-scale combinatorial optimisation problems potentially addressable by QAOA-like or annealing-based quantum optimisation heuristics . Recent studies on healthcare digital twins suggest that incorporating such quantum-enabled modules could expand the dimensionality and coupling that can be handled in patient-specific models, although practical QDT deployments in clinical settings remain at a very early stage .
2. Biological System
For broader biological systems (metabolic networks, cellular communities, ecosystemâorganism interactions), QDTs could target the computational core of stochastic reactionâdiffusion models and network dynamics. For example, analysing gene-regulatory or signalling networks often reduces to dynamics on large, sparse graphs and to sampling from high-dimensional stationary distributions, which might benefit from quantum walk-based algorithms or quantum-accelerated sampling schemes. Single-cell omics data, on the other hand, naturally lead to very high-dimensional, sparse data analysis problems in clustering and trajectory inference, where QML-based representation learning could, in principle, provide more expressive embeddings. Early work on hybrid quantumâclassical digital twins with explicit uncertainty quantification suggests that quantum-enhanced modules may be useful for propagating uncertainty through such complex networked models , although concrete large-scale biological QDT deployments are not yet available.
3. Aerospace
Aerospace systems (aircraft, satellites, launch vehicles) include coupled structural, thermal, fluid-dynamic and control subsystems, stringent safety requirements and real-time decision support. In this setting, QDTs can, in principle, enhance several classes of tasks. Trajectory optimisation, constellation design and mission planning give rise to large discrete or mixed-integer optimisation problems, natural candidates for QAOA-type or annealing-based quantum optimisation modules . High-fidelity aerothermodynamic and structural simulations involve solving very large sparse linear systems or PDEs, for which quantum linear solvers and quantum PDE algorithms promise asymptotic speedups in certain regimes . However, bridging these theoretical advantages to industry-grade meshes, complex boundary conditions and certification constraints remains a long-term challenge. Finally, predictive maintenance and fault diagnosis in avionics and structural health monitoring could exploit QML models trained offline on telemetry and sensor streams, with distilled classical surrogates deployed in safety-critical, real-time DT instances. Emerging aerospace-oriented QDT concepts emphasise real-time optimisation and autonomous decision making in space applications, but so far remain largely at the roadmap or prototype stage .
4. Smart City Smart city DTs integrate transport, energy, buildings, environment and infrastructure at urban scale. The resulting workloads combine network flow optimisation (for traffic and logistics), unit commitment and demand response (for energy systems), and multi-agent coordination (for mobility and services). Quantum optimisation algorithms could be used to tackle large combinatorial subproblems such as traffic signal timing plans, demand-response scheduling or distributed energy resource dispatch within hybrid planning frameworks . QML models for spatio-temporal forecasting may aid in predicting traffic, energy demand and environmental conditions under uncertainty, improving the robustness of city-level DT predictions. Moreover, quantum-safe cryptography and, in selected settings, quantum key distribution can be integrated into smart-city QDT architectures to secure twin-to-twin and cross-organisational communication links involving critical infrastructure data, complementing conventional cybersecurity mechanisms.
5. Mobility and Transportation
In transportation (vehicle fleets, network flows, autonomous mobility), digital twins support routing, real-time prediction and safety-critical decision-making. QDTs may contribute quantum-enabled combinatorial optimisation for vehicle routing, crew and charging scheduling, and empty-vehicle repositioning problems, many of which can be expressed as large QUBO instances amenable to QAOA or annealing heuristics . For autonomous and connected vehicles, quantum reinforcement learning and planning methods might be explored for complex multi-agent decision-making under uncertainty, albeit primarily in offline training and evaluation loops rather than latency-critical control. Large-scale scenario simulation and rare-event analysis, potentially accelerated by quantum Monte Carlo techniques, could further enhance the predictive power of mobility DTs.
6. Smart Manufacturing
Manufacturing is among the most mature DT domains. Integrating quantum computing yields several concrete opportunities. Production planning, job-shop and flow-shop scheduling, and dynamic reconfiguration of manufacturing cells are classical NP-hard optimisation problems that can be encoded as QUBO or Ising models and targeted by QAOA-style or annealing-based quantum optimisation modules . On the analytics side, QML models can be trained on high-dimensional process data, vibration and image streams for predictive maintenance and quality monitoring. Industrial collaborations between Bosch and Multiverse Computing, for instance, have begun to investigate quantum and quantum-inspired algorithms as accelerators within manufacturing DT simulation workflows, including prototype QDTs for factory processes . These efforts remain at the proof-of-concept stage, but they illustrate how quantum modules can be integrated into existing industrial DT pipelines.
7. Robotics
For robotics (collaborative robots, autonomous drones/vehicles), digital twins must model kinematics, dynamics, sensing, environment interactions and control loops under tight real-time constraints. Quantum enhancement is therefore more realistic in computationally heavy but latency-tolerant components such as motion planning, task allocation and policy optimisation, rather than in low-level servo control. For example, multi-robot task allocation and path planning can be formulated as combinatorial optimisation problems for QAOA or annealing-based solvers, while QRL or quantum-enhanced policy search methods may be explored for learning control policies in complex, partially observed environments. In all cases, quantum modules are likely to operate offline or in supervisory loops, with policies or planners ultimately executed by classical controllers in the physical robots.
8. Natural Systems and Environment
Environmental systems (climate, atmosphere, water/earth systems, pollution dispersion, ecosystem dynamics) are inherently high-dimensional, stochastic and multi-scale. DTs in this domain rely heavily on numerical solution of PDEs, stochastic differential equations and large sparse linear systems, as well as Monte Carlo sampling for uncertainty quantification and extreme-event analysis. Quantum algorithms for linear systems and PDEs, and quantum Monte Carlo via amplitude estimation, offer asymptotic advantages for some of these tasks . In practice, near-term applications are more likely to involve reduced-order or surrogate models whose most expensive components are offloaded to quantum accelerators. A recent quantum-DT project in green hydrogen plant optimisation exemplifies a process-level environmental QDT, where quantum optimisation is explored for improving operational efficiency under stochastic renewable input and demand conditions .
9. Agriculture
Agricultural digital twins model soilâcropâclimate coupling, water and nutrient flows, crop growth and farm-operation scheduling, typically over large spatial domains and long time horizons. The resulting decision problems (e.g., irrigation and fertiliser scheduling, machinery deployment, crop rotation planning) are high-dimensional, stochastic and subject to multiple, sometimes conflicting objectives (yield, cost, environmental impact). Quantum optimisation could be used to explore large combinatorial decision spaces in multi-objective planning under constraints, while QML models may help fuse satellite remote-sensing, in-field IoT and weather data to estimate latent soilâcrop states and to forecast yields under uncertainty. Although concrete QDT case studies in agriculture are currently rare, recent work on QMLâDT integration and quantum-enhanced uncertainty handling in supply-chain settings suggests that similar hybrid quantumâclassical architectures could benefit agricultural planning and logistics .
10. Commerce
In commerce, digital twins simulate supply chains, asset lifecycles, customer behaviour and risk. QDTs here naturally build on quantum optimisation and QML modules. Network design, facility location, inventory positioning and routing can be formulated as large-scale combinatorial optimisation problems suitable for QAOA-type or annealing-based quantum heuristics, possibly embedded in multi-stage stochastic optimisation frameworks. For risk and uncertainty management, recent work has proposed hybrid quantumâclassical supply-chain digital twins in which quantum feature transformations and variational classifiers are used to improve demand forecasting, anomaly detection and uncertainty propagation, with the DT serving as a testbed for policy evaluation . These studies report potential gains in forecast accuracy and computational efficiency on small-scale testbeds, but emphasise that scalability and hardware noise remain significant challenges. Quantum-safe cryptography and secure multiparty computation techniques can further complement QDT-based financial and supply-chain systems, although these are orthogonal to the core simulation and optimisation roles of quantum computing. Quantumâclassical hybrid DTs for uncertainty quantification in noisy quantum devices themselves represent another line of work, highlighting the use of QDT concepts to analyse quantum hardware performance .
11. Education and Training
Educational and training digital twins (virtual labs, immersive learning, complex system simulators) can benefit from QDTs in two complementary ways. First, DTs that emulate complex physical or engineered systems can incorporate quantum-accelerated simulation modulesâfor example, in quantum physics, chemistry or materials scienceâto provide learners with interactive access to phenomena that are otherwise computationally or experimentally inaccessible . Second, DTs of the learning process itself may employ QML/QAI models to construct personalised learning paths and adaptive training scenarios based on rich learner interaction data, with the DT providing a sandbox for testing pedagogical policies before deployment. At present, concrete QDT deployments in education and training remain largely conceptual, but the combination of DTs with QML/QAI aligns with broader trends in data-driven, simulation-based education.
Open Challenges and Future Directions
While AI-empowered digital twins have demonstrated transformative potential across diverse domains, several fundamental challenges remain unresolved. This section identifies critical open problems and outlines promising research directions that require concerted efforts from the research community.
Bridging Physics and AI.  Despite progress in physics-informed neural networks (PINNs) and hybrid modeling, a fundamental tension persists between data-driven AI and physics-based simulations . Multi-scale integration remains challengingâcardiovascular digital twins, for instance, must seamlessly model from molecular biochemistry to organ-level hemodynamics while maintaining computational efficiency . Uncertainty quantification in hybrid models lacks principled methods for combining uncertainties from both physics-based and data-driven components, critical for high-stakes applications like aerospace and healthcare. Ensuring AI models preserve physical constraints such as conservation laws and causality during long-horizon predictions calls for architectures with embedded physical priors, including Hamiltonian neural networks and symplectic integrators.
Scalability and Real-Time Performance.  As digital twin systems expand from single assets to interconnected networks spanning factories, supply chains, or cities, computational challenges grow exponentially . Hierarchical and federated architectures are needed to orchestrate distributed twins while ensuring data consistency and minimizing communication overhead. Neural surrogate models provide speedups for simulation, but must balance computational efficiency and physical accuracy. Real-time adaptation to physical system changes, including degradation and failure, remains difficult. Edge-deployed models require efficient compression techniques that preserve physical consistency while meeting latency constraints .
Trustworthiness and Ethics.  Digital twins increasingly inform safety-critical decisions, demanding rigorous assurance of transparency, robustness, and fairness . Explainability should extend beyond feature attribution toward causal and physically grounded interpretations. Counterfactual reasoning and uncertainty quantification can support âwhat-ifâ scenario analysis, improving diagnostic reliability. Robustness to adversarial perturbations and cascading feedback failures remains an open problem. Formal verification techniques adapted from control theory could help certify system safety. Ethical governance frameworks that define accountability, auditability, and privacy protection are essential to ensure trust and societal acceptance .
Human-AI Collaboration.  Human oversight remains indispensable in digital twin ecosystems, especially for complex, uncertain, or high-stakes contexts. Effective collaboration requires intuitive interfaces where natural language queries are translated into simulation or optimization actions. Immersive visualization (VR/AR) and multimodal communication can make high-dimensional predictions interpretable to operators. Human-in-the-loop reinforcement learning and shared autonomy frameworks can integrate expert feedback dynamically, enabling systems that learn user preferences and calibrate trust levels through transparent uncertainty communication.
Standardization and Cross-Domain Transfer.  The absence of universal data models, ontologies, and communication protocols limits interoperability between digital twin platforms . Community-wide standardizationâled by efforts like the Digital Twin Consortiumâis necessary to promote modular, interoperable architectures. Transfer learning offers an opportunity to reuse knowledge from mature industrial domains to accelerate deployment in emerging areas such as synthetic biology. Benchmark datasets covering diverse operational and failure modes, together with unified evaluation metrics balancing accuracy, robustness, and interpretability, are vital for reproducible research.
Emerging Frontiers.  New frontiers for digital twins extend into global and human-scale systems. Climate, epidemic, and socio-economic digital twins demand integration of physical models with human behavioral and policy dimensions . Whole-human twins and quantum-level modeling represent long-term goals requiring breakthroughs in multi-scale data integration, biological modeling, and computational efficiency. Extending digital twin concepts to social systems raises profound questions about human agency, reflexivity, and ethics, underscoring the need for interdisciplinary collaboration among AI researchers, social scientists, and ethicists.
The convergence of AI and digital twin technologies is reshaping how we perceive, predict, and manage complex physical systems. Realizing this potential will require bridging physics and AI, scaling architectures for real-time autonomy, ensuring ethical governance, and fostering human-centered collaboration. We envision a future in which AI-powered digital twins serve as cognitive infrastructuresâmonitoring, reasoning, and adapting across scales from molecular to planetaryâdriving a new era of intelligent, sustainable, and trustworthy system management.
Acknowledgement
This work was in part supported by Lehigh Universityâs CORE and RIG grants. In addition, we gratefully acknowledge Ruoxi Chen for her assistance in the early stages of figure preparation and for insightful discussions during her time as a visiting student at Lehigh University.
đ ë ŒëŹž ìê°ìëŁ (Figures)









