Abstract
Digital twins bring value to mechanical and aerospace systems by speeding up development, reducing risk, predicting issues and reducing sustainment costs. Realizing these benefits at scale requires a structured and intentional approach to digital twin conception, design, development, operation and sustainment. To bring maximal value, a digital twin does not need to be an exquisite virtual replica but instead must be envisioned to be fit for purpose, where the determination of fitness depends on the capability needs and the cost–benefit trade-offs.
Similar content being viewed by others
Shell buckling for programmable metafluids
Phase-change memory via a phase-changeable self-confined nano-filament
Optimal blade pitch control for enhanced vertical-axis wind turbine performance
Main
A digital twin has been defined as “a set of virtual information constructs that mimics the structure, context, and behavior of an individual or unique physical asset, is dynamically updated with data from its physical twin throughout its life cycle, and ultimately informs decisions that realize value”1. This definition highlights the ways in which a digital twin goes beyond traditional modeling and simulation to bring value, through the bidirectional feedback flows between the physical asset and its virtual representation2. The flow of data from physical to virtual enables the virtual representation to be dynamically updated, thus tailoring it to the specific behavior of its physical counterpart, while the flow from virtual to physical induces changes in the physical asset, through control, sensor steering or other manipulations of the physical system. The virtual representation typically comprises a disparate set of models representing different disciplines, subsystems and components of the physical system. These models may be based on an understanding of the physical principles (for instance, analytical and empirical models that are grounded in physical governing laws). These models may also be purely numerical, such as statistical models that fit from data and machine learning. These models may also be hybrid combinations of physics-based and data-driven models. These digital twin elements and the centrality of the bidirectional flows to a digital twin are depicted in Fig. 1. The figure also depicts the role of a human in the loop and the importance of holistic and continual validation, verification and uncertainty quantification.
Decades of advances in modeling and simulation are powerful precursors to digital twins. A digital twin builds on these advances but goes beyond them, as a capability to provide a physical asset with a unique digital identity. Underlying this unique identity are the requirements to digitally capture relevant asset aspects, such as actual geometry, realized performance and other physical state observations. Also required is the ability to present and manipulate the digital twin view to humans for complex decision-making. These aspects may be achieved by a data-centric approach that measures and stores asset-specific information, or by a model-centric approach that combines analysis, models and data from other indirect measurements. These analyses can leverage well-known simulation technologies, such as behavioral simulation, computational fluid dynamics, finite-element analysis and computer-aided design, combined with additional numerical methods to twin the model with the physical asset.
Digital twin use cases
Notably, across aerospace and mechanical engineering, the use cases for digital twins are vast and the potential benefits large. Digital twins have the potential to speed up development, reduce risk, predict issues and drive reduced sustainment costs. Digital twins enable new ways to collaborate across the life cycle and the supply chain. In the design phase, digital twins provide an opportunity to unlock the advantages of digital engineering, including reduced prototype cycle times, lowered technical risks and reduced experimental test costs. In manufacturing, digital twins could enable improved first-time yield, products optimized with design-for-manufacturing considerations, and optimized factory operations to improve cycle time and cost. In operations, digital twins can lead to improved system capability, increased operational availability, reduced maintenance costs and reduced root-cause-corrective-action turnaround times.
Multiple use cases are currently under investigation or development in each of these life-cycle phases.
Design and engineering
In the engineering phase, we differentiate the models of a product or a critical component (models largely developed in recent decades) from digital twins when data from current or historical mission, manufacturing or testing are available to tailor predictions to a specific physical asset or an aggregation of assets. The combination of data with models enables a drastic enhancement of the accuracy of the predicted information to directly support decisions related to the modeled asset(s).
The main use cases in this phase are: (1) exchange digital twins across the supply chain to accelerate the design cycle time; (2) mission analysis-based design, leveraging mission data to run tailored simulations to improve performance; (3) perform design optimization to improve the quality, performance or cost of the design; (4) virtual verification and validation, performing system and component testing in the digital space before any physical activities take place to reduce cost; and (5) virtual certification, leveraging digital twin validation to receive certification credit and reduce certification cost.
Manufacturing
In the manufacturing phase, there are two different groups of digital twins: those related to the product and those related to the factory optimization (such as manufacturing process, supply chain and logistics). The latter are covered by ref. 3.
Prototype digital twins for products can be used to convey to manufacturing, from engineering, the design of the system or the part. They represent the parts in terms of geometry, tolerances and material information (called model-based definition) and enable the automation of manufacturing and inspection processes with the objective to improve quality and to reduce the time for ramping up production. When the inspection of the part occurs and the part assumes a unique identity (for instance, based on serialization), then the data tailor the prototype digital twin to the precise physical counterpart (for instance, tolerances are removed and effective dimensions are used). These digital twins are also used in other use cases, such as product acceptance (do the products satisfy tolerances?) or additional analysis for overall part suitability to improve production yield. For example, a virtual assembly is performed using digital twins of the parts to verify that the tolerances stack up in a compliant and consistent way, or to enable the operator to perform the activities in an augmented or virtual reality environment with the objective to reduce cost and improve quality of assembly.
Sustainment and operation
One use of digital twins is to perform mission and maintenance training; this capability improves mission readiness and reduces human errors with clear benefits for operators. In addition, many valuable decisions in this area can be improved with the support of digital twins. An example is the decision to remove from the wing an engine that is currently in operation. A digital twin providing the estimation of the lifing based on historical missions enables the operator to execute this decision when most appropriate. The decision is based on the actual condition of the engine, instead of being scheduled at a fixed time. These condition-based maintenance and prognostic use cases are improving the availability of critical products in the field.
Another relevant use case in this area is to employ digital twins to improve mission performance. If the digital twin provides real-time estimation of key aspects in fractions of the decision time, it can be evaluated multiple times to perform a ‘what if’ analysis during operation and support more robust decisions, or be used to analyze whether more performance can be provided for a limited amount of time. As a concrete example, drawn from ref. 4, the latest update of the Pratt & Whitney F119 turbofan has digital twins of critical components, which enable the estimation of the remaining life based on the historical usage of the engine. The information from the digital twins shows that the parts of some engines, due to a use pattern that is different from the designed one, have a longer life than expected. This condition enables saving a substantial amount of maintenance cost—about US$800 million across the aircraft’s life cycle—if the maintenance is delayed to later in time, or if the additional margin is used to improve the performance in the remaining scheduled life.
The digital twin as an asset in its own right
Investment in digital twin technologies is required to realize the potential benefits across all these life-cycle phases and achieve the full promise of digital twins. These benefits can be realized at scale only if the digital twin comes to be thought of as an asset in its own right, an asset that—just like the physical asset—must be conceptualized, architected, designed, built, deployed and sustained (Fig. 2). Each one of these phases of the digital twin life cycle requires trade-offs that weigh digital twin development and operational costs with the value added.
In the following, we summarize some of the key considerations needed in the different phases of the digital twin life cycle. Our experience is that current practice is to build the digital twin after the physical asset has been realized, cobbling together whatever models and data flows are available. While this may provide sufficient value in some settings, for complex engineering systems—especially those with long service lives—we emphasize the importance of a structured and intentional approach to the life cycle of the digital twin, with investments paralleling the investments that are made over the life cycle of a physical asset. This systematic life-cycle approach to a digital twin remains largely absent in current practices.
Digital twin conception
Just as the conceptual design phase is critical in the design of a physical asset, realizing the full value of a digital twin requires an intentional conceptual design phase for the digital twin itself. This conception phase should define requirements and architecture for the digital twin, including identifying its operational environment(s), potential use cases and the associated fidelity requirements. The digital twin’s conceptual design phase should identify existing models, existing data sources and gaps in digital twin capabilities (including gaps in data, models, integration and workflows) that will require development to fill. In conceiving the digital twin and defining its requirements, it is also essential to architect a plan for investment and sustainment of the digital twin over the full life cycle of the physical counterpart. Just as manufacturability and maintainability are considerations in the design of a physical asset, they must be considerations in the design of a digital twin.
Digital twin development
Development of the digital twin requires model development, software engineering, integration and workflow establishment. In some cases, the virtual representation underlying the digital twin may use models employed for the design of the physical asset, while in other cases, the digital twin may need its own models to be developed. Surrogate modeling may play a particularly important role in developing the digital twin, due to the need to satisfy stringent computational constraints in operation, particularly for digital twins that support real-time updating and decision support. Surrogate modeling provides a way to balance model fidelity with computational costs, where the surrogates may be derived via physics-based reduced-order modeling, empirical data fits, machine learning or hybrid combinations thereof.
Digital twin validation
Validation is “the process of determining the degree to which a model is an accurate representation of the real world from the perspective of the intended uses of the model”5,6. The latter part of this definition, emphasizing the ‘intended uses of the model’ is highly relevant for digital twin validation. As described in the National Academies of Sciences, Engineering, and Medicine (NASEM) consensus study report2, a digital twin must be fit for purpose, meaning that it must balance the fidelity of predictions, as defined in the digital twin’s conceptual design phase, with the computational demands (such as time, power and complexity). The notion of being fit for purpose must also factor in acceptable levels of uncertainty to drive decisions and risk preferences of decision-makers. Figure 2 highlights two distinct roles of validation—the substantial validation effort that must accompany digital twin development and the continual validation that must accompany digital twin updates. Methodologies to achieve continual in-the-loop validation largely remain an open challenge, yet represent a key enabler to realizing the potential of digital twins for high-consequence and safety-critical applications.
Digital twin operation
In the operation phase, the digital twin is evaluated to support the decision-making process. It is composed of two major aspects: (1) a digital framework to execute, evaluate and visualize the digital twin; and (2) business processes that integrate the digital twin information for decision-making.
The digital framework collects data from the different authoritative sources, such as sensor data and other context information or product configurations, and provides the storage and computational services to execute the digital twin. Digital twin execution entails running the needed simulations and, given the user queries or requests, providing visualization of the information. Digital twin execution may also entail feeding a more complex set of services for information composition or autonomous decision-making.
The business processes are heavily dependent on the particular use case; their description is outside the scope of this Perspective.
Digital twin updating
Updating is a critical element of a digital twin and a key distinguishing factor that enables a digital twin to bring added value. Digital twin updating may take place with varying extents and over different timescales. Whatever the extent and timescale of the updates, the notion of continuous validation is critical. Updates that take place in real time or near real time (for instance, via data assimilation or parameter-estimation algorithms) must be accompanied by a form of certification that the components of the digital twin virtual representation remain within acceptable bounds. For example, one could imagine a scenario where the physical counterpart undergoes a change in state, so that the acquired data lead to a parameter update that puts the digital twin’s underlying models beyond their validation conditions. Such a scenario requires action to recognize that the digital twin’s predictions are no longer trusted, triggering a more relevant update that entails additional validation and possibly additional digital twin development (such as the development of a new or refined model). This is similarly important for digital twin updates that take place on slower timescales. For example, if the physical counterpart is updated (such as through hardware or software upgrades, part replacement and so forth), one needs to ensure that the linkage between the physical counterpart and the digital twin is maintained. In current practices, these kinds of updates are largely human driven and manual. A further aspect of digital twin updating is the updating of the digital twin virtual representation itself, not because the physical counterpart has changed but because there is a pull to update the digital twin’s capabilities. This pull could be because the operational environment and its associated fidelity requirements have changed, because the availability and/or quality of data have changed, or because there has been an advance in modeling or algorithmic capability that would bring value to the digital twin (for instance, through reduced computational time or reduced power consumption for onboard processing). Ensuring that the digital twin’s virtual representation receives its own maintenance, sustainment and upgrades is critical to realizing the full value, as discussed in the NASEM report (Ch. 7 in ref. 2).
Digital twin disposal
The disposal of a digital twin must follow the end of life of the physical assets. There are two major scenarios:
-
The physical asset is at the end of life and appropriately disposed. In this case, the digital twin, depending on the regulations and policies, will have to be archived for future analysis or for historical data valuable for new generation of similar products.
-
The physical asset, or a part of it, is having a second life (such as in a circular economy). In this specific case, the digital twin will become a relevant source of information for the reuse of the assets or of the parts. For example, in the second life of a battery, the digital twin is still very relevant for the forecast of the battery’s storage capacity.
An identical twin or a fit-for-purpose capability?
It is tempting to imagine a digital twin as a true ‘identical twin’ of a physical asset, representing the system at an exquisite level of detail and providing a visual representation that looks and acts just like the physical asset. Indeed, some definitions of digital twins promote this vision. While this vision may be realizable in limited cases for simple systems with narrow requirement sets, for complex mechanical and aerospace systems it is our view that it is neither beneficial nor tractable to envision a digital twin in this way. Instead, a digital twin must be envisioned to be fit for purpose, where the determination of fitness depends on the capability needs and the cost–benefit trade-offs. For complex mechanical and aerospace systems, there is value in conceiving a digital twin to have a capability progression. For example, for the same physical asset, one could imagine a digital twin at different levels of capability such as design and prototyping, monitoring and diagnostics, forecasting and predictive analytics, control, autonomous operations, and system-of-systems optimization. Each one of these capability levels places different demands on the digital twin. While the trade-offs will necessarily be case specific, we discuss in the following some of the key trade-off elements that should be considered.
Requirements and trade-offs in fit-for-purpose digital twins
In the conception and development of a digital twin, the digital twin solution must meet multiple business and technical requirements spanning across the entire life cycle. For example, the cost of operation, including the continual alignment of the digital twin with its physical counterpart, depends heavily on the amount of data and computation needed to carry out a trustable forecast of a specific aspect of the physical asset. In some cases, these data and computational requirements may not be directly aligned with business objectives. In such cases, the indirect benefit offered by the digital twin must be factored into resource allocation decisions.
The typical business requirements can be summarized as follows: (1) the return on investments (meaning, the value delivered during the operation minus the cost of development, operation and disposal) must be acceptable to the business; (2) the digital twin forecast must be trustable to allow improvement of business decisions; (3) the user experience must meet user expectations; and (4) the digital twin must be integrated within the business processes (for instance, operator workflows) of the physical asset.
Digital twins for mechanical and aerospace engineering systems must be consistent with the same first principles of the physical twins. This implies that, in the exploration of the solution space, one of the most important trade-offs to be analyzed is the availability of models and the complexity of modeling based on physics-based knowledge versus the amount of data needed for training and continual updating. At one extreme, we may have detailed models that capture (or abstract) the entire physics of the asset and require a minimal set of data, over time, to become tuned to the real world as an accurate digital replica. At the other extreme, we may have data-driven models (such as machine learning models) without any a priori physics or behavioral knowledge that are trained on collected data and that learn the input–output relations based only on data. In the middle ground, we may have a set of hybrid techniques, such as physics-informed machine learning7, data-driven reduced modeling8,9, sparse identification of nonlinear dynamical systems10 and scientific machine learning11, which use a mix of physics-based information and field data to build a digital twin that better fits the purpose and matches the business and technical requirements.
Figure 3 shows, in the vertical direction, the potential trade-offs between physics-based models and data, and the impact of key requirements. For example, physics-based models are capable of forecasting even with limited historical information, while data-driven models are limited when extrapolating outside of the supplied data. The chosen trade-off will dictate the satisfaction of technical requirements, such as: (1) frequency and latency of evaluation of the digital twin; (2) accuracy over time and update frequency; (3) visualization and query capabilities; and (4) data collection and computational requirements during development, validation and operation.
A fundamental requirement to be satisfied by the digital twin solution is the capability to maintain, across the entire life cycle, the association of the digital identity with the physical asset and its configuration, such that the digital twin is always representing the correct asset. Techniques for a unique digital signature of a mechanical asset can be based on unique physical characteristics or can rely on manual processes. We note that manual processes are prone to errors and not always compatible with business processes during operation.
When moving into implementation, the digital twin must also satisfy architectural and deployment requirements. The related operational framework is a key aspect that, if not already present, needs to be conceived and designed. In some autonomous or real-time aerospace systems, access to cloud infrastructures is infeasible and the entire operational process (from data acquisition to computation, forecast and action) must be executed on edge or embedded devices. In contrast, if a cloud infrastructure is adopted, the communication bandwidth and latency requirements, from the field to the cloud, must be met to enable the flow of data with the required frequency. These deployment requirements drastically limit the space of possible digital twin solutions. An ever-increasing concern that deserves a deeper analysis is the cybersecurity of the entire solution: altering the outputs or the data inputs of the digital twin may have severe business and safety consequences; as such, security requirements must be part of the digital twin design from its initial conception.
The critical role of uncertainty quantification in a digital twin
As mentioned in the previous section, the business trust of the digital twin forecast is one of the most critical requirements to be demonstrated. Currently, this is tackled with a pragmatic approach of extensive testing and system validation. While this may be acceptable and feasible for many use cases, for others a more formal approach to conduct uncertainty quantification of the forecast based on robust proof is needed. Uncertainty quantification via formalized methods, such as the probabilistic graphical modeling approach12, implies considerable modeling, simulation and testing effort; in current practices, formal uncertainty quantification is typically not conducted until the end of the design phase. Sensitivity analysis is a powerful yet underutilized aspect of uncertainty quantification. Instead, preference is given to collect more and more data, even though doing so does not necessarily improve the accuracy and trust of the digital twin. The incorporation of more formal uncertainty quantification approaches, including sensitivity analysis, is an important direction for future digital twin development. There are important standardization efforts in this area13, aligning terminologies across stakeholders and providing common frameworks for uncertainty quantification. This is an important area of research for which future progress is needed to increase trust and enable a larger deployment of digital twin solutions, especially for safety-critical systems.
Final remarks
Digital twins have already been demonstrated to bring value in mechanical and aerospace systems, speeding up development, reducing risk, predicting issues and driving reduced sustainment costs. While early digital twin deployments have been successful, realizing the full potential benefits at scale requires a more structured and intentional approach to digital twin conception, design, development, operation and sustainment. We advocate for the value of considering the digital twin as an asset in its own right, and approaching its development with the same principles of cost–benefit–risk trade-offs that are employed in the design and development of physical assets. While we have focused on digital twins in mechanical and aerospace engineering, this Perspective is also highly relevant to digital twins in other domains, where a similarly intentional approach to digital twin development would prove valuable. Lastly, we note that a discussion of workforce gaps and opportunities is beyond the scope of this Perspective, but remains a critically important consideration in achieving the full benefits of digital twins.
References
-
AIAA Digital Engineering Integration Committee Digital Twin: Definition and Value AIAA and AIA Position Paper (AIAA, 2020).
-
Foundational Research Gaps and Future Directions for Digital Twins Consensus Study Report (NASEM, 2023).
-
Tao, F., Zhang, H. & Zhang, C. Advancements and challenges of digital twins in industry. Nat. Comput. Sci. https://doi.org/10.1038/s43588-024-00603-w (2024).
-
Trimble, S. Digital twin helps Pratt expand F119 performance for F-22. Aviation Week and Space Technology (3 October 2022).
-
NRC Assessing the Reliability of Complex Models: Mathematical and Statistical Foundations of Verification, Validation, and Uncertainty Quantification (National Academies Press, 2012).
-
AIAA Computational Fluid Dynamics Committee Guide for the Verification and Validation of Computational Fluid Dynamics Simulations (AIAA, 1998).
-
Karniadakis, G. E. et al. Physics-informed machine learning. Nat. Rev. Phys. 3, 422–440 (2021).
Google Scholar
-
Peherstorfer, B. & Willcox, K. Data-driven operator inference for nonintrusive projection-based model reduction. Comput. Methods Appl. Mech. Eng. 306, 196–215 (2016).
Google Scholar
-
Ghattas, O. & Willcox, K. Learning physics-based models from data: perspectives from inverse problems and model reduction. Acta Numer. 30, 445–554 (2021).
Google Scholar
-
Brunton, S. L., Proctor, J. L. & Kutz, J. N. Discovering governing equations from data by sparse identification of nonlinear dynamical systems. Proc. Natl Acad. Sci. USA 113, 3932–3937 (2016).
Google Scholar
-
Baker, N. et al. Workshop Report on Basic Research Needs for Scientific Machine Learning: Core Technologies for Artificial Intelligence (US DOE Office of Science, 2019).
-
Kapteyn, M. G., Pretorius, J. & Willcox, K. A probabilistic graphical model foundation for enabling predictive digital twins at scale. Nat. Comput. Sci. 1, 337–347 (2021).
Google Scholar
-
Verification, Validation, and Uncertainty Quantification Terminology in Computational Modeling and Simulation Standard (ASME, 2022).
Acknowledgements
K.W. acknowledges support from Department of Energy grant DE-SC0021239 and Air Force Office of Scientific Research grants FA9550-21-1-0084 and FA9550-22-1-0419.
Author information
Authors and Affiliations
Contributions
Both authors contributed to the conception, writing and editing of this paper.
Corresponding author
Ethics declarations
Competing interests
A.F. is part of RTX, an aerospace and defense enterprise company. K.W. declares no competing interests.
Peer review
Peer review information
Nature Computational Science thanks Kristi Morgansen and Omer San for their contribution to the peer review of this work. Primary Handling Editor: Fernando Chirigati, in collaboration with the Nature Computational Science team.
Additional information
Publisher’s note Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
Reprints and permissions
About this article
Cite this article
Ferrari, A., Willcox, K. Digital twins in mechanical and aerospace engineering.
Nat Comput Sci 4, 178–183 (2024). https://doi.org/10.1038/s43588-024-00613-8
-
Received: 22 January 2024
-
Accepted: 20 February 2024
-
Published: 26 March 2024
-
Issue Date: March 2024
-
DOI: https://doi.org/10.1038/s43588-024-00613-8