1. Introduction
As the Internet of Things (IoT) and Big Data gain widespread adoption, Digital Twin technology is surging in popularity. According to a recent study by Markets and Markets[
1], the Digital Twins (DTs) market, valued at
$6.9 billion in 2022, is projected to soar to a staggering
$73.5 billion by 2027, demonstrating an impressive compound annual growth rate of 60.6% over a five-year span. The study identifies key industries poised to heavily invest in this technology, including Automotive & Transportation, Energy & Utilities, Infrastructure, Aerospace, Healthcare, and Oil & Gas. These industries are leveraging DTs in various applications, such as product design and development, predictive maintenance, performance monitoring, supply chain management, and business optimization. While larger organizations have established a robust pathway for adopting Digital Twins, there remains significant uncertainty regarding the speed and effectiveness with which Small and Medium Enterprises (SMEs) can embrace this innovative approach. In response to this challenge, the European community has initiated several programs, such as IoTwins, Change2Twin, and DigitBrain, [
2,
3,
4] with the goal of facilitating a swift and cost-effective adoption of Digital Twins among small enterprises.
IoTwins is a project supported by the European Union under the H2020 program, Commencing in September 2019, IoTwins successfully concluded its planned activities by August 2022. IoTwins exploits the big data available in Industry 4.0 to devise smarter and more effective approaches to predictive maintenance and operation optimization in industrial manufacturing. Likewise, IoTwins wants to leverage big data to derive descriptive insights about the operations and processes developed in facilities like buildings, smart power grids, data centers, etc. On the basis of such descriptive information, optimization techniques can provide efficient facility management plans, operation optimal schedules, and renovation/maintenance plans.
The primary objective of the IoTwins project is to reduce the technological barriers faced by SMEs when seeking to create intelligent digital services that enhance their industrial production capabilities. Aligned with the guiding principles of the fourth industrial revolution (I4.0), IoTwins aims to support the digital transition of industrial factories by providing them with a methodology and tools that leverage the potential of the DT computing paradigm and of ICT technologies such as Cloud/Edge computing, Big Data and Machine Learning (ML), yet hiding away from the beneficiary the complexity of adopting such technologies separately and together. The paper contribution are summarized as:
Definition of a hybrid and distributed DT model;
Design of a DT reference architecture inspired to the RAMI 4.0 model[
5];
Implementation of an open platform that adheres to the architecture’s principles;
Definition of a set of guidelines for the agile development of industrial DTs.
The paper structure is as follows. in
Section 2, we introduce the background and discuss some related work. In
Section 3, we provide a definition of the hybrid and distributed DT model devised within the IoTwins project.
Section 4 is devoted to the description of the RAMI-inspired IoTwins reference architecture. In
Section 5, we discuss some implementation details of the IoTwins platform, while in
Section 6 a practical implementation and deployment of DT-based application is discussed. Finally,
Section 7 concludes the work.
2. Background and Related work
The idea of the DT was formulated by Michael Grieves and introduced back in 2003 at the University of Michigan [
6]. Since its inception, this concept has gained growing prominence due to its capacity to generate a digital representation of a tangible entity. Initially, the adoption of DT gained traction within the manufacturing sector, followed by its expansion into the realms of the Internet of Things (IoT) and cyber-physical systems (CPSs)[
7]. Furthermore, it has captured the attention of various technical communities and professionals spanning diverse industries. These stakeholders have identified shared aspects between their existing methodologies, concepts, and needs. Consequently, the concept of digital twins has undergone adaptation and augmentation, resulting in its multifaceted interpretation shaped by the specific domain and intended utility. The literature is full of attempts to give a formal and exhaustive definition of DT [
8,
9,
10,
11,
12]. In this article, we will stick to the definition proposed by [
13], as we believe it embodies the concept of DT that best fits our aim: "A DT is a comprehensive digital representation of an individual product. It includes the properties, conditions, and behavior(s) of the real-life object through models and data. A DT is a set of realistic models that can simulate an object’s behavior in the deployed environment. The DT represents and reflects its physical twin and remains its virtual counterpart across the object’s entire lifecycle".
More specifically,
Industrial Digital Twins(IDTs) refer to virtual representations of physical industrial assets, processes, and systems. An IDT is essentially a dynamic, real-time digital representation of a physical asset or process, which allows for monitoring, analysis, and optimization in a virtual environment. For a more exhaustive list of definitions of IDTs the reader may refer to [
14]. Examples of industries that benefit from IDTs include manufacturing, energy (power plants, oil and gas facilities), transportation (aircraft, trains, vehicles), healthcare (patient monitoring and treatment optimization), and more. Here is a non-exhaustive list of potential areas of application of IDTs in industrial scenarios:
Real-time monitoring and data collection. IDTs continuously gather data from sensors and other sources in the real world. This data is then used to update the virtual representation, ensuring that it closely reflects the current state of the physical asset or process.
Predictive analysis. By using historical data and advanced analytics, IDTs can predict future behavior and potential issues. This enables proactive maintenance and reduces downtime.
Remote operation and control. With a IDT, operators can remotely monitor and control physical assets, even in challenging or hazardous environments. This is particularly useful for industries like energy, oil and gas, and manufacturing.
Optimization and testing. IDTs allow for testing different scenarios and configurations in a virtual environment before applying changes to the actual physical asset. This can lead to optimized processes and reduced trial-and-error efforts.
Reduced downtime and maintenance Costs. By predicting and preventing issues before they occur, IDTs can help reduce unplanned downtime and maintenance costs.
Lifecycle management: IDTs cover the entire lifecycle of an asset, from design and development to operation and maintenance, and even decommissioning.
Authors of [
14] list the enabling technologies that turns the DT paradigm into a concrete opportunity for industries to undertake the digitization process fostered by the Industry 4.0 revolution. First and foremost, advancements in
data acquisition and analysis technologies (e.g., advanced wireless networks, communication protocols, big data analytics) enable the build of faithful digital representations and strong integration of the latter with their physical counterparts [
15,
16].
High-fidelity modeling comprises both accurate interpretation of raw data from the physical asset into knowledge and the fusion of information generated by the virtual model targeting at the optimization of the physical entity [
17]. In that respect, AI is frequently employed to create models that utilize established inputs and outputs acquired from the real-world system, aiding in the comprehension of how physical properties interact with each other [
18].
Simulation stands out as a significant enabler of the DTs, primarily due to the enhanced value it brings by enabling seamless real-time communication between virtual and physical assets. Simulating DT behavior results to significant opportunities with respect to the mutual optimization of the virtual and the physical model in terms of operation and maintenance schedule [
19].
In the aim of establishing a
comprehensive framework for constructing DTs, various endeavors have emerged to devise distinct modeling approaches. These efforts encompass the formulation of diverse DT models categorized by methodological tiers, known as layers, which facilitate the seamless exchange of information between the physical and virtual domains. Among the different modeling methods, the five-layer structure, six-layer structure, three-step process, and five-dimensional modeling are being commonly found in literature [
20]. In [
21], authors develop a DT reference model and architecture, and apply them in an industrial case. Inspired by the RAMI 4.0 reference model[
5], they propose a layered model that comprises of three dimensions coordinate that describes all crucial aspects of DTs, namely architecture, value life cycle and integration. Efforts to device a DT development methodology can be found in the literature. [
22] proposes a methodology design using model-driven engineering (MDE)that strives toward being both flexible and generic. According to this approach a DT is first modeled as a composition of basic components that provide basic functionalities (e.g., identification, storage, communication, security, etc.); second, an aggregated DT is defined as a hierarchical composition of other DTs. A generic reference architecture based on these concepts and a concrete implementation methodology are proposed using AutomationML[
23].
This paper presents a DT model, a reference architecture and a development methodology to help companies, both big and SME, leverage the potential of the DT paradigm to boost their digitization process. The proposed DT model embodies all the state-of-the-art approaches, i.e., the data-driven, the model-driven and the hybrid ones, and accounts for the possibility of having the DT distributed over the industrial cloud continuum (OT-to-Cloud). Similarly to [
21], our reference IT architecture inspires to the principles of the RAMI 4.0 architecture; besides that, we propose an implementation of a software prototype of a DT-based PaaS for concrete use by the developers. Finally, we deliver a methodology and a set of practical guidelines for the implementation and TOSCA-based deployment of complex, containerized and distributed DTs in the computing continuum.
3. Design of a hybrid and distributed Digital Twins model for industrial scenarios
The digital transition fostered by the Industry 4.0 revolution is a process that companies need to undertake in order to survive the tough challenges posed by the global market. Aligned with this view, the European H2020 project
IoTwins1 aims to support European SMEs to undertake the digital transformation by "democratizing" the access to the most prominent information technologies. The main objective pursued by IoTwins is to lower the technological barriers encountered by SMEs in the intent of adopting Big Data-driven intelligent services that will help them extract knowledge from the daily collected data and exploit it to achieve better business performance. In the practice, IoTwins wants to deliver an open-software platform and a toolbox that manufacturers can harness to easily develop and operate Big Data-fueled, AI-powered and Cloud/Edge-enabled industrial applications.
Among all enabling technologies called upon by IoTwins, the Digital Twins (DTs) paradigm plays a pivotal role. As a European
Innovation Action project
2, IoTwins employs several industrial pilots to investigate opportunities and issues of adopting DTs (as well as of other technologies) by industrial players operating in the fields of manufacturing and facility management. IoTwins intends to also push further on the uptake of such a powerful paradigm by proposing a scalable DT model that can be easily replicated in several other industrial settings as well as verticals. IoTwins envisions a strong involvement of its industrial partners (IPs) in the definition of the DTs model. IPs are requested to provide the test-bed facility to support the DT model validation and the domain expertise to define the business requirements for the model design and implementation. The IoTwins test-beds are divided into three categories: i)
manufacturing test-beds, ii)
facility/infrastructure management test-beds, and iii) test-beds for in-field verification of the
replicability, scalability, and standardization of the proposed approach, as well as the generation of new business models. In particular, four industrial pilots in the manufacturing sector are aimed at providing predictive maintenance services that use sensor data to forecast the time to failure and produce maintenance plans that optimize maintenance costs. Three large-scale test-beds concerning facility management cover online monitoring and optimization in IT facilities and smart grids, as well as intervention planning and infrastructure maintenance/renovation on sport facilities based on data collected by sophisticated and heterogeneous monitoring infrastructures. The five final test-beds, on the other hand, have been carefully chosen to demonstrate the replicability of the proposed IoTwins methodology in diverse industries, the scalability of the adopted solutions, and their ability to assist SMEs in developing new business models. In
Table 1, the twelve IoTwins test-beds are listed along with a synthetic description of their claimed objectives.
The DT model developed by IoTwins exploits big data and domain expert knowledge to accurately represent a complex system (i.e., industrial plant, process, facility), along with its application-relevant performance indicators, with the ambition of being capable of correctly predicting the temporal evolution and dynamics of the system. From a conceptual perspective, IoTwins envisions the development of the following DT types: simulation-based, leveraging either agent-oriented modelling or physical modelling; data-driven, exploiting the most recent ML/DL techniques; hybrid, combining the power of physics and data modeling. A sample hybrid DT was developed in Test-bed 3. Here, in the aim of developing a predictive model of a machine part’s faults, a ML model was implemented and trained with real-time data gathered from the sensors on board on the machine part. Given the scarcity of data sensed when the component ran in a near-to-faulty mode, the manufacturing company had to resort to simulation in order to generate synthetic data related to the malfunctioning of the part. Such data were then used along with real data to train a robust ML model for the prediction purpose. Furthermore, in order to cater for the computing needs of factories, as well as for the strict requirements of certain types of industrial applications, IoTwins fosters a hierarchical distribution and interworking of DTs that includes: i) IoT Twins, featuring lightweight models of specific components and performing big-data processing and local control for quality management operations (low latency and high reliability); Edge Twins, deployed at plant gateways, providing higher level control knobs and orchestrating Internet of Things (IoT) sensors and actuators in a production locality, thus fostering local optimizations and interoperability; Cloud Twins, performing time-consuming and typically off-line parallel simulation and deep-learning, feeding the edge twin with pre-elaborated predictive models to be efficiently executed at the premises of production plants for monitoring/control/tuning purposes.
From a methodological perspective, the whole DT model definition process envisaged two iterations: in the first iteration, requirements elicited from the industrial manufacturing and facility management IPs (say "group A") were used to design a first draft of the DT model, which was then prototyped in collaboration with the scientific partners and validated on the testbeds owned by group A IPs; in the second iteration, results collected from the mentioned experiments and new requirements elicited from the so-called replicability IPs (say "group B") contributed to refining the DT model, which was eventually validated on a second set of testbeds owned by group B IPs.
Finally, in the
Figure 1, we depict a graphical representation of the the hybrid and distributed DT model developed by the IoTwins project. First, a DT (depicted as a blue-filled box) is a distributed entity that may span the whole continuum of an industrial scenario, ranging from the remote Cloud to the factory premises, where it can run on Edge nodes and on field devices/PLC (labled as "IoT")
3. Second, operating a distributed DT requires a well-designed, robust and scalable data backbone that will have to support the exchange of signals among the DT components (depicted as red hollow arrows) and shipping a lot of data from the field to the Cloud (red-filled arrows). Finally, three types of models are allowed in this scenario: agent-based simulation, physics model simulation and trained ML models. A DT can belong to one of this category, or can be a hybrid implementation that mixes the simulation and the data-driven approach to achieve its goal (as is the case of Test-bed 3).
4. A RAMI-inspired reference architecture
The
IoTwins architecture was designed to support the DT model discussed in
Section 3. It draws inspiration from the Reference Architectural Model Industry 4.0 (RAMI 4.0)[
5] developed by the German Electrical and Electronic Manufacturers’ Association (ZVEI) to support Industry 4.0 initiatives. RAMI 4.0 provides a unified model that ensures all the stakeholders involved in an I4.0 ecosystem to share data and information in an efficient and effective way. The RAMI 4.0 model comprises three "axes" named Life Cycle value stream, Hierarchy levels and Architecture Layers respectively. Grounding on the IEC62890 standard
4, the Life cycle value stream axis provides a view of the product life cycle from conception to disposal. The foundation of the Hierarchy levels axis are IEC62264
5 and IEC61512
6 respectively, which aim to represent different functional levels of a factory. Finally, the Architecture layers axis enables the transformation of industrial assets into their interoperable Digital Twins. This axis support most of IoTwins research objectives. In the following, we report a short description of it.
The Architecture layers axis defines a framework where the physical world meet the digital one and a strong interconnection among the manufacturing operations is enabled. A layered view of the axis is depicted in the left-end of
Figure 2. At the bottom, the
Assets layer identifies and describes the real assets in the physical world. It comprises sensors, devices, machine parts, machines, machine groups, etc. The
Integration layer describes the digital equivalents of physical assets. This layer is where the transition from the physical world to the cyber space begins. The
Communications layer addresses mechanisms for the interoperable exchange of information between digital assets. The
Information layer defines data services such as provisioning and integration that can be leveraged to exchange data among functions, services, and components. The
Functional layer provides the runtime and modelling environment to build functions and services to support the business. Finally, the
Business layer defines organizational and business-related applications, processes and operations. Inspired by the RAMI4.0 Architecture layers axis, the
IoTwins architecture aims to provide a reference architecture to guide the implementation of software platforms for building, operating and maintaining DT-based industrial applications. IoTwins proposes a logical, layered architecture defining the functions that the software platform will have to offer. In the
Figure 2, we illustrate the IoTwins architecture[
2,
24] and highlight the mapping between the architecture functions and the RAMI4.0 concepts that the each function addresses.
The IoTwins architecture address RAMI concepts ranging from the ’Integration’ to the ’Functional’. The RAMI ’Asset’ layer represents physical things in the shop floor (production line machines, sensors, actuators, etc.) that need to be connected to the digital world; therefore, being IoTwins a reference architecture for software platforms, it will build on top of those assets. Similarly to the RAMI4.0 model, the IoTwins architecture adopts a layered approach, with each layer leveraging the functionalities provided by the lower layer and offering services to the upper layer. In the following, a detailed description of such layers and the RAMI4.0 concepts they address is given.
Runtime Layer.
In order to project an industrial physical object ("thing", in the following) into the digital world, it is necessary to set up a computing environment where the digital alter ego of the thing can live. This layer is responsible for abstracting the available computing resources (that may range from very small IoT devices to large High Performance Computing(HPC) clusters) and providing a virtualized execution environment that can flexible accommodate the computational demand of DTs. Technologies that may serve the mentioned purpose include lightweight virtualization, hypervisor-based virtualization and HPC middleware.
Resource Layer.
We refer to a "resource" as a virtual computing entity that can be activated on demand on the virtualized execution environment. A very simple form of a digital copy of a physical thing can even be implemented by means of a simple resource (e.g., a microservice mirroring a sensor’s data). This layer recommends a set of services that aim to guarantee full dependability of resources when they are operational. Among others, this layer takes care of tracking all running resources, scheduling new resources on demand, monitoring the resources status and implementing resource resilience strategies.
Application Layer.
This layer includes application templates and toolkits to assist developers in implementing the DT business logic. In this regard, two approaches are supported for the the development of DTs: a data-driven approach, which makes intensive use of ML/DL and Big Data analytics, and a model-driven approach, which relies on the use of software simulation. The support for a combined use of data-driven and model-driven techniques for the development of a hybrid DT is also provided.
Authentication and Authorization.
This is the ingress point to access the services offered in the Platform layer and in the Application layer. Here, access procedures are put in force in order to grant a safe and controlled access to both data and services.
Data security.
Privacy is a strong requirement that cross-cuts all the layers. Since private and sensitive data may be handled, both raw data coming from the shop floor and those elaborated along the path must be secured. In that respect, procedures to protect data at rest (e.g., data anonymization, data encryption) as well as data in transit (e.g., secure communication channels) must be enforced.
5. The IoTwins platform: a software prototype
We implemented a software platform that adheres to the design principles specified by the IoTwins reference architecture. As mentioned in
Section 3, the platform went through two refinement iterations governed by an overarching process of i) requirements elicitation from industrial testbeds and ii) use case validation. In this section, we will disclose some implementation details of the platform prototype, illustrating the software environment that we instrumented to operate industrial Digital Twins in the computing continuum. Finally, we will discuss the point of view of software developers taking advantage of the tools offered by the IoTwins platform to build and deploy a DT-based application.
5.1. IoTwins Platform’s implementation details
To support the coding of the platform’s prototype, many commonly available and highly mature open-source software products have been used. Considering the maturity level of such software, whose Technology Readiness Level (TRL) is in the range 8 - 9, and the extensive tests that all industrial partners ran to attain their goals, we can claim that the final version of the platform released at the end of the project achieved an estimated maturity level of TRL 6.
The platform is a distributed system made up of a number of software components that can be deployed in the continuum. Despite the architectural design addresses three level of computing environments, namely IoT, Edge and Cloud, for space reason in the following we give a description of the platform components that are commonly deployed on the Cloud side and on the Edge side, as depicted in the
Figure 3 and
Figure 4 respectively. In the final tests conducted at the end of the project, the Cloud-side platform components were deployed in the private data center of one of the technology providers that participated in the project. For what concern the Edge components, they were deployed on a commodity PC equipped with the Linux Ubuntu 20 OS. Bottom-up, we are going to briefly discuss the software products that provide an implementation of the functions/services populating the hierarchical layers of the RAMI-inspired IoTwins architecture. In each figure, a coloured legend recalls which RAMI’s layer a given platform component addresses.
At the bottom of the
Figure 3 the software tools implementing the RAMI’s integration layer on the Cloud side are depicted. Virtualization software like the Openstack
7, Docker
8 and Kubernetes
9 was employed to abstract the underlying physical computing and storage resources and offer them as a pool of virtual resources that can be managed in an easier and more uniform way. The heart of the platform is the INDIGO Orchestrator [
25], which contributes to implement services belonging to the RAMI’s functional layer. The INDIGO Orchestrator is a TOSCA-compliant[
26] cloud orchestrator in charge of accepting application deployment requests, scheduling virtual computing resources and enforcing provisioning workflows that serve the requests. Beside fulfilling the application deployment task, INDIGO is capable of enforcing actions that guarantee the scalability and fault-tolerance of the deployed applications. In the next section, a sample provisioning workflow enforced by the INDIGO orchestrator is described. The platform provides services to help the developer implement the data backbone of their DTs. On the left-end of the figure, the RabbitMQ message broker and a set of streaming data adapters are depicted in yellow, signifying their belonging to the RAMI’s Communication layer. Those components are responsible for the gathering of data from the Edge and their adaptation to the application’s required format. On the right-end of the picture, some DBMS tools (MinIO
10 as object-storage, InfluxDB
11 as time-series, MongoDB
12 as NoSQL) are depicted along with the adapters that developers can craft with the support of the telegraf
13 tool. The support for data management is then enriched with two open repositories that store Docker-based modules that can be re-used for DT development purpose. Finally, in the figure a sample of potential industrial applications is depicted:
Control logic is a software application that implements the business logic of an industrial control loop;
Data pre-processor is a component that filters/adapts streamed data before feeding them to the control logic;
ML training is a neural network that needs to be trained both on locally-stored historical data and on real-time data streamed from the Edge.
In the
Figure 4, the software components implementing the Edge side of the IoTwins platform are depicted. In the bottom, the virtualization tools Apache Mesos
14, Marathon
15 and Docker implement the integration layer prescribed by the RAMI architecture. Despite the Mesos tool is designed to virtualize and manage a cluster of computing nodes, it perfectly accomplishes the management duties of just one node. Furthermore, in future developments it will cope fine with scenarios where multiple Edge nodes need to be managed. Tools like Marathon and Chronos, in their turn, will offer the developer the opportunity of running long-running and job-like computing instances respectively. The reader may have noticed that there is no orchestrator component deployed in the Edge. The reason is that, at design time, we decided to centralize the orchestration functionality in the Cloud, so there is just one component (the INDIGO orchestrator, indeed) responsible for orchestrating the computing resources belonging to the Cloud/Edge continuum. On the Edge side, orchestration instructions are remotely triggered by the INDIGO orchestrator to the Mesos tool via the REST interface. Similarly to the Cloud deployment, software adapters are provided to meet both data stream and data storage adaptation needs. On the left end of the picture a
Data Collector component is displayed. It will cater for the need of collecting data from IoT devices in the field independently of the communication protocol they use. Finally, a list of sample components is depicted in green that represent potential applications that the developer may decide to run on the Edge.
5.2. Digital Twins implementation guidelines
As mentioned in
Section 3, the IoTwins project aimed at "democratizing" access to most advanced information technology, i.e., providing SMEs with cheap instruments to embark the digital transformation facilitated by I4.0. The IoTwins open platform is one of the most remarkable project outcomes, benefiting both the research community and industry. The platform not only allows users to easily build DT-based services to support their own businesses; it also enables service developers to implement new, re-usable software modules that will contribute to the consolidation of an ecosystem of freely accessible and composable services from which the community can draw to accelerate the development path of their DTs. To support this objective, the IoTwins initiative created an open service repository and populated it with some representative and popular services. It also created a set of rules for service developers to follow in order to design platform-compatible DTs.
In the IoTwins framework, a DT can be a very simple software module, running in either the Cloud or the Edge, or can take the form of a complex chain of interworking modules deployed and running along the continuum. Because the Docker framework is used as containerization technology in both Cloud and Edge-level runtime environments, each IoTwins DT must be developed as a (composition of) Docker container(s). This enables "coding" the DT once and deploying it anywhere along the Docker-powered Cloud/Edge continuum. Furthermore, in order to meet the INDIGO’s "orchestrability" criteria, an ad-hoc TOSCA template including instructions on how to deploy the software module must be provided for each DT. We will use the term "Toskerization" to designate to the process of creating a new DT, which stems from a crasis between the phrases TOSCA and Docker.
A Toskerized DT is a software bundle that embeds a Dockerized image of the service (i.e., a service that can run in a Docker runtime environment) and a TOSCA file (indeed, the service template) that instructs the INDIGO PaaS Orchestrator on how to correctly deploy the service upon the user request. The Dockerized service can be built out of a plain Docker image publicly retrievable from any of the available Docker repositories, by optionally adding extra layers according to the specific needs. Basically, the Dockerized DT must be configured to accept a list of input parameters, that the user may want to pass for correctly configuring the service, and a list of output parameters needed to configure other Dockerized services that might be deployed along in a service chain fashion. The described approach simplifies the service deployment operations in a consistent way. In fact, the user does not have to directly handle the software package/tool that they need, nor they have to manipulate configuration files. The declarative approach offered by the TOSCA standard (and enforced by the INDIGO orchestrator component of the IoTwins platform) offers the user an easy way to declare what their deployment objectives are and takes care of the entire deployment process (i.e., pulling the software tools from a repository, install it in a private computing space, configure it, re-run the deployment in case of temporary failures, etc.). Service developers who want to make their DTs orchestrable by the IoTwins platform must follow a few simple principles that show how the Toskerization process works. The process includes the following steps:
Creating the Docker image. The developer will have to explore public Docker repositories to search for existing dockerized images of the service that they wish to implement. If such an image is not available, they will have to build the image from scratch. When editing the Dockerfile, the developer will have to make sure that the image accepts input values for the correct configuration of the service at runtime: the easiest way to accomplish this goal is to pass input data values through C-shell environment’s variables, as most ready-to-use docker images are already set to read variables from the environment where they execute; in the case that further configuration work is needed, the developer will have to create ad-hoc scripts to be injected in the Docker image and run them
Uploading the Docker image to the IoTwins repository. The platform is provided with a private docker container repository that offer storing, image retrieval and text-based search functionalities. Once uploaded on the repository, the docker image can be accessed by the orchestrator in a transparent way in order to enforce provisioning tasks
Coding the TOSCA template. The TOSCA standard offers a declarative approach to define the topology and the provisioning workflow of cloud-based and distributed application. The developer is in charge of mastering the TOSCA-compliant blueprint containing the instructions to provision the software modules implementing the DT. Instructions are declarative statements concerning, e.g., the computing capacity requested by the DT, the configuration properties of the software modules (docker components) that the DT is composed of, their mutual dependencies, etc.
Testing the DT orchestrability. In order to run functional tests on the mastered TOSCA template and on the related provisioned services, the developer can make use of two front-end tools: a command line interface (CLI) and a web-based interface. Both tools let the developer send deployment commands to the INDIGO orchestrator and monitor/debug the provisioning process in a sandbox environment.
6. Building and provisioning an industrial Digital Twin
We present an illustrative case study that delves into the definition, implementation and deployment of a DT adhering to the IoTwins reference model and spanning the Cloud/Edge industrial continuum. The case under discussion is a tangible instance of a developmental initiative conducted within the framework of an IoTwins testbed implementation. The collaborating partner owning the testbed expressed the need of developing an AI-driven application for identifying irregularities within an industrial machine tool during its production process. In pursuit of this objective, the machine was outfitted with sensors designed to capture specific physical metrics (such as load, forces, vibrations, etc.). The intention was to accumulate an extensive dataset, which would subsequently be used to train a ML model proficient in detecting potential functional anomalies associated with the aforementioned machine tool. Due to the time-critical nature of the control loop (detecting potential anomalies quickly reduces the risk of tool damage), it is essential to execute the trained ML model as proximately as feasible to the data sources. While an Edge computing node satisfies this demand, it cannot ensure the computational power necessary for ML model training. In contrast, the Cloud emerges as the more suitable computing environment, offering the requested capacity for training the ML model. Developers may encounter various technical and administrative challenges when implementing such an application. To begin, they must establish a data path across the continuum (from sensors through Edge to Cloud) to ensure a continuous flow of data to both the ML models - the one under training and the trained one. Secondly, the modules constituting the DT applications need appropriate configuration, interconnection, and deployment within a distributed computing environment. Subsequently, once the Cloud-based ML model has been successfully trained on data at rest, it must be transferred to the Edge where it will receive real-time data.
The IoTwins platform equips developers with tools and services to confront these challenges and expedite application development. Following the guidelines outlined in in the previous section, developers will explore the platform’s repository to identify reusable modules that align with the application’s objectives. Fortunately, the repository provides a range of Docker containers for constructing the DT’s data backbone. These encompass message brokers for data distribution, adaptable connectors for data format and protocol conversion, as well as databases for diverse data storage requirements. These software components can be effortlessly assembled by developers to create the desired data path. Developers are tasked with implementing the ML model and any supplementary elements pertinent to the application’s business logic. These components should be containerized using the Docker framework and uploaded to the repository. When all DT components are available in Dockerized forms, developers will focus on mastering the TOSCA blueprint that governs the entire DT structure. This involves populating the blueprint with configuration parameters for the components, instructions for component interconnections, and the callback mechanism necessary for migrating the ML model component from Cloud to Edge following its training. An excerpt from the TOSCA blueprint, demonstrating the interdependencies between two components of the DT data infrastructure (specifically, the message broker and a connector), is displayed in Listing 1.
Subsequently, the developer will submit the TOSCA blueprint to the orchestrator, which is responsible for deploying the DT components based on the provided instructions. The orchestrator has the capability to deduce the correct order of deployments to be executed. In
Figure 5, we present a visual representation of the DT components deployed within the cloud continuum. In this illustration, reused components are denoted in a deep blue color, while components that the developer crafted from scratch are shown in a light blue hue. The solid lines indicate the exchange of data between pairs of components, while the dashed line represents the migration path of the ML model. Some components are deployed on an Edge node within the factory premises, while others are provisioned in the Cloud. The data backbone supporting the application’s logic comprises several key components:
Listing 1. Exert of a TOSCA blueprint for the provisioning and wiring of a RabbitMQ docker instance and a Telegraf connector instance |
|
The data generated in the field is first collected by the message broker and then routed to its intended destinations through these components. On the Cloud side, the ML model component, depicted as the "Anomaly Detection (AD) Model," retrieves data from the MinIO object storage. This data is essential for the training process of the ML model, which is a critical part of anomaly detection. On the Edge end, the trained AD model component subscribes to the message broker in order to receive fresh data generated by the sources. Based on those data, it will have to detect potential anomalies and notify with the Alarm component. Finally, the Data Polisher filter and cleans in-transit data before those get to the Cloud.
7. Conclusions
This paper provides an overview of Digital Twins reference model as presented in the EU-funded IoTwins project. In particular, the paper focuses on the design and development of a software prototype of an open platform to support the agile implementation of Digital Twins-based applications in industrial settings. A set of development guidelines is also delivered that will help developers to build Digital Twins by composing existing containerized software. This work contributes to the state of the art in the field by proposing a RAMI-inspired Digital Twins reference architecture and delivering a novel and practical "build-by-compose" approach and easy-to-use tools to development of industrial applications in the industrial continuum, i.e., the environment spanning the whole chain of computing resources ranging from the shop floor to the remote cloud. The positive outcomes attained in IoTwins, encompassing improved time-to.market of digital twin applications and decreased financial investment, mark a substantial advancement. We claim these achievements will contribute to motivate SMEs to expedite the digitazation processes fostered by Industry 4.0.
Abbreviations
The following abbreviations are used in this manuscript:
DTs |
Digital Twins |
IDTs |
Industrial Digital Twins |
IoT |
Internet of Things |
RAMI |
Reference Architectural Model Industry 4.0 |
I4.0 |
Industry4.0 |
SME |
Small and Medium enterprise |
PLC |
Programmable Logic Controller |
TRL |
Technology Readiness Level |
DBMS |
Data Base Management System |
References
- Markets and Markets. Digital Twin Market. https://www.marketsandmarkets.com/Market-Reports/digital-twin-market-225269522.html. last seen on May 10, 2023.
- Costantini, A.; Di Modica, G.; Ahouangonou, J.C.; Duma, D.C.; Martelli, B.; Galletti, M.; Antonacci, M.; Nehls, D.; Bellavista, P.; Delamarre, C.; Cesini, D. IoTwins: Toward Implementation of Distributed Digital Twins in Industry 4.0 Settings. MDPI Computers 2022, 11. [Google Scholar] [CrossRef]
- The Change2Twins consortium. The H2020 Change2Twins project. https://www.change2twin.eu/. last seen on May 10, 2023.
- The DIGITbrain consortium. The H2020 DIGITbrain project. https://digitbrain.eu/. last seen on May 10, 2023.
- RAMI 4.0 - ISA.
- Grieves, M. Product Lifecycle Management: Driving the Next Generation of Lean Thinking; McGraw-Hill, 2005. [Google Scholar]
- Minerva, R.; Lee, G.M.; Crespi, N. Digital Twin in the IoT Context: A Survey on Technical Features, Scenarios, and Architectural Models. Proceedings of the IEEE 2020, 108, 1785–1824. [Google Scholar] [CrossRef]
- Miller, A.M.; Alvarez, R.; Hartman, N. Towards an extended model-based definition for the digital twin. Computer-Aided Design and Applications 2018, 15, 880–891. [Google Scholar] [CrossRef]
- Okita, T.; Kawabata, T.; Murayama, H.; Nishino, N.; Aichi, M. A new concept of digital twin of artifact systems: synthesizing monitoring/inspections, physical/numerical models, and social system models. Procedia CIRP 2019, 79, 667–672. [Google Scholar] [CrossRef]
- Boschert, S.; Heinrich, C.; Rosen, R. Next Generation Digital Twin. Proceedings - 12th International Symposium on Tools and Methods of Competitive Engineering, TMCE 2018, 2018.
- Van Os, J. The Digital Twin throughout the Lifecycle. SNAME Maritime Convention, 2018.
- Shafto, M.; Conroy, M.; Doyle, R.; Gleassgen, E.; Kemp, C.; LeMoigne, J.; Wang, L. Draft modelling, simulation, information technology & processing roadmap. Technology Area 2010, 11. [Google Scholar]
- Haag, S.; Anderl, R. Digital twin – Proof of concept. Manufacturing letters 2018, 15, 64–66. [Google Scholar] [CrossRef]
- Stavropoulos, P.; Mourtzis, D. Chapter 10 - Digital twins in industry 4.0. In Design and Operation of Production Networks for Mass Personalization in the Era of Cloud Technology; Elsevier, 2022; pp. 277–316.
- Mourtzis, D. Simulation in the design and operation of manufacturing systems: state of the art and new trends. International Journal of Production Research 2020, 58, 1927–1949. [Google Scholar] [CrossRef]
- Fuller, A.; Fan, Z.; Day, C.; Barlow, C. Digital twin: Enabling technologies, challenges and open research. IEEE access 2020, 8, 108952–108971. [Google Scholar] [CrossRef]
- Tao, F.; Qi, Q.; Liu, A.; Kusiak, A. Data-driven smart manufacturing. Journal of Manufacturing Systems 2018, 48, 157–169. [Google Scholar] [CrossRef]
- Liu, M.; Fang, S.; Dong, H.; Xu, C. Review of digital twin about concepts, technologies, and industrial applications. Journal of Manufacturing Systems 2021, 58, 346–361. [Google Scholar] [CrossRef]
- Madni, A.M.; Madni, C.C.; Lucero, S.D. Leveraging digital twin technology in model-based systems engineering. Systems 2019, 7, 7. [Google Scholar] [CrossRef]
- Zhang, H.; Ma, L.; Sun, J.; Lin, H.; Thürer, M. Digital twin in services and industrial product service systems:: Review and analysis. Procedia CIRP 2019, 83, 57–60. [Google Scholar] [CrossRef]
- Aheleroff, S.; Xu, X.; Zhong, R.Y.; Lu, Y. Digital Twin as a Service (DTaaS) in Industry 4.0: An Architecture Reference Model. Advanced Engineering Informatics 2021, 47, 101225. [Google Scholar] [CrossRef]
- Schroeder, G.N.; Steinmetz, C.; Rodrigues, R.N.; Henriques, R.V.B.; Rettberg, A.; Pereira, C.E. A Methodology for Digital Twin Modeling and Deployment for Industry 4.0. Proceedings of the IEEE 2021, 109, 556–567. [Google Scholar] [CrossRef]
- Drath, R.; Luder, A.; Peschke, J.; Hundt, L. AutomationML-the glue for seamless automation engineering. 2008 IEEE International Conference on Emerging Technologies and Factory Automation. IEEE, 2008, pp. 616–623.
- Borghesi, A.; Di Modica, G.; Bellavista, P.; Gowtham, V.; Willner, A.; Nehls, D.; Kintzler, F.; Cejka, S.; Tisbeni, S.R.; Costantini, A.; Galletti, M.; Antonacci, M.; Ahouangonou, J.C. IoTwins: Design and implementation of a platform for the management of digital twins in industrial scenarios. Proceedings - 21st IEEE/ACM International Symposium on Cluster, Cloud and Internet Computing, CCGrid 2021, 2021, pp. 625–633.
- INDIGO DataCloud consortium. INDIGO-DataCloud Project. https://repo.indigo-datacloud.eu/. Last accessed: Sep 4th, 2023.
- OASIS. TOSCA Simple Profile in YAML Version 1.3. http://docs.oasis-open.org/tosca/TOSCA-Simple-Profile-YAML/v1.3/TOSCA-Simple-Profile-YAML-v1.3.html. Last accessed: May 10, 2023.
1 |
|
2 |
Projects of this kind enforce activities of prototyping, piloting and market replication. |
3 |
Some samples of concrete DT-based applications have been depicted in the figure as red-filled boxes with a blu outline |
4 |
|
5 |
|
6 |
|
7 |
|
8 |
|
9 |
|
10 |
|
11 |
|
12 |
|
13 |
|
14 |
|
15 |
|
|
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).