Next Article in Journal
Design and Optimization of a Contour Deep Loosening and Stubble Pressing Mechanism Suitable for Strip Tillage
Previous Article in Journal
Prioritizing Early-Stage Start-Up Investment Alternatives Under Uncertainty: A Venture Capital Perspective
Previous Article in Special Issue
Research on Digital Restoration and Innovative Utilization of Taohuawu Woodblock New Year Prints Based on Edge Detection and Color Clustering
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Knowledge Graphs and Artificial Intelligence for the Implementation of Cognitive Heritage Digital Twins

1
VAST LAB, PIN, Piazza dell’Università 1, 59100 Prato, Italy
2
ARIADNE Research Infrastructure AISBL, Rue de Grand-Bigard 14, Berchen-Sainte-Agathe, 1082 Bruxelles, Belgium
*
Author to whom correspondence should be addressed.
Appl. Sci. 2025, 15(18), 10061; https://doi.org/10.3390/app151810061
Submission received: 19 July 2025 / Revised: 30 August 2025 / Accepted: 11 September 2025 / Published: 15 September 2025

Abstract

This paper explores the integration of Artificial Intelligence and semantic technologies to support the creation of intelligent Heritage Digital Twins, digital constructs capable of representing, interpreting, and reasoning over cultural data. This study focuses on transforming the often fragmented and unstructured documentation produced in cultural heritage into coherent Knowledge Graphs aligned with internationally recognised standards and ontologies. Two complementary AI-assisted workflows are proposed: one for extracting and formalising structured knowledge from heritage science reports and another for enhancing AI models through the integration of curated ontological knowledge. The experiments demonstrate how this synergy facilitates both the retrieval and the reuse of complex information while ensuring interpretability and semantic consistency. Beyond technical efficacy, this paper also addresses the ethical implications of AI use in cultural heritage, with particular attention to transparency, bias mitigation, and meaningful representation of diverse narratives. The results highlight the importance of a reflexive and ethically grounded deployment of AI, where knowledge extraction and machine learning are guided by structured ontologies and human oversight, to ensure conceptual rigour and respect for cultural complexity.

1. Introduction

It is said that Helen of Troy, the most beautiful woman in the world according to Greek mythology, never set foot on the shores of Asia Minor. In a dramatic and unsettling version of the myth, told by Euripides and echoed in later traditions, it was not the real Helen whom Paris carried off to Troy. Instead, a simulacrum, an eidolon, a perfect likeness devoid of substance, was sent by the gods to take her place (this version of the myth is attested in Apollodorus, Epitome III. 5; Euripides, Electra, 128 and Helen 31 ff.; Servius on Virgil’s Aeneid, I. 655 and II. 595; Stesichorus, quoted by Tzetzes, On Lycophron 113). While the true Helen waited in silence in Egypt, perhaps unaware of the war unfolding in her name, it was her image that ignited desire, deceived kings, and plunged the world into its most storied conflict [1]. The eidolon was not a lie but a form with agency: it acted, provoked, and transformed the course of history. Though lacking life of its own, it possessed real effectiveness.
In this figure of the operative double, of an image that does not merely represent but intervenes, there is a striking anticipation of the modern notion of digital twin. Such an entity is designed not just to mirror the original but to accompany or replace it in action, to operate in its stead, often with greater reach and precision. A digital twin, to fulfil this role, must do more than resemble. Like Helen’s phantom, it must be endowed with structured intelligence and a “grammar” that sustains its coherence. The story of Helen’s eidolon speaks to an ancient intuition: that a likeness, when crafted with precision and animated with purpose, can operate in the world with consequences equal to, or even surpassing, those of its original. This mythic logic finds a contemporary counterpart in the digital age, where cultural entities are increasingly represented by what might be called simulacra, i.e., constructed doubles whose very name shares etymological roots with the verb to simulate. These digital twins are not inert copies but dynamic, structured representations designed to engage with the world through processes of analysis, interaction, and prediction.
In the realm of cultural heritage, the digital twin has evolved far beyond the notion of a static visualisation or isolated model. It now constitutes a complex informational organism, integrating diverse layers of documentation: historical texts, scientific analyses, sensory data, and curatorial knowledge. To preserve coherence across this multiplicity of sources, ontologies provide the foundational structure to build the memory of the digital twin by serving as its semantic architecture and allowing knowledge to be expressed formally, queried effectively, and shared across systems and institutions.
Yet, structure alone cannot suffice. Just as Helen’s eidolon required a divine breath to become operative, digital twins must be endowed with intelligence in order to act meaningfully. Artificial Intelligence, when guided by ontological frameworks, enables digital twins to transcend passive representation. Through the semantic interpretation of textual and visual sources, the detection of patterns in environmental data, and the simulation of complex behaviours, AI seems ideal to become the interpretive engine of the twin. In this framework, natural language processing, entity recognition, semantic inference, and predictive modelling are no longer isolated functionalities but become part of a coherent system of knowledge and memory production.
A distinctive innovative aspect of this study is the systematic combination of Artificial Intelligence and ontologies in the construction of digital twins. Previous approaches, especially in cultural heritage, have often relied on AI without semantic guarantees, or on ontological models without operational integration. The framework presented here links the two dimensions inseparably: ontologies act as the semantic backbone that ensures transparency, coherence, and interoperability, while AI functions as the cognitive engine that extracts, enriches, and interprets knowledge from the complex body of documentation that sustains the digital twin, including analytical reports, curatorial records, and other heritage data. It is precisely in the synergy between ontologies and AI that the true potential of digital twins can unfold. Their mutual interaction constitutes the epistemological core of a new generation of digital twins in which ontologies ensure that AI models are not only efficient but also transparent, interpretable, and anchored in shared conceptual frameworks. In return, AI can enrich semantic information by identifying latent connections, disambiguating meanings, and populating knowledge graphs with contextually valid information. The result is a self-expanding semantic system, capable of both reasoning and remembering, a system that learns, explains, and operates in alignment with the epistemic and ethical values of cultural heritage research.
This paper is structured as follows: Section 2 outlines the historical evolution of digital twins, particularly those enhanced by AI, from their origins in engineering and aerospace to their gradual redefinition within the cultural heritage domain. Section 3 introduces the Heritage Digital Twin and its associated ontology, designed to implement cognitively enriched, conceptually expressive constructs. Section 4 explores the uses of new Artificial Intelligence technologies to extract knowledge from cultural heritage documentation and encode it according to the semantics of ontological models widely used in this domain to build the knowledge graphs of Heritage Digital Twins. Section 5 investigates how knowledge graphs can be used to enrich the semantic toolbox of Artificial Intelligences, transforming them into agents capable of reasoning on large and intricate amounts of data and answering complex scientific questions formulated in natural language. Section 6 presents a case study where the entire pipeline is tested on a set of scientific reports from the ARTEMIS project documentation to identify and distil information used to populate the knowledge graphs that serve as the engine of the Heritage Digital Twins and is used to enrich and reinforce AI components, enhancing their ability to support advanced functionalities such as natural language interfaces, intelligent querying, and adaptive interaction with users. Following the discussion in Section 7, which highlights the potential of these technologies while also highlighting the inherent risks in their application, particularly when not subject to rigorous human oversight, this paper moves towards Section 8, offering some conclusive notes and future perspectives.

2. Digital Twins: From Conceptual Models to Operational Systems

The relevance of digital twins has steadily expanded across disciplines, reflecting their growing role as sophisticated digital counterparts of real-world entities and systems [2,3]. While often loosely applied to any virtual replica, the term also designates complex, structured information systems capable of interfacing with the physical world. Initially developed to enhance industrial performance, optimise processes, and anticipate failures, digital twins have since evolved to support various domains for their ability to improve planning decisions [4], especially by integrating multi-source data [5].
The origins of digital twins are often traced to NASA’s simulation technologies in the 1960s, which allowed engineers to replicate space conditions virtually—an epistemological necessity in the face of inaccessibility [6]. Although the term “digital twin” formally emerged only in 2010 [7], the conceptual lineage is deeper and more complex, as is often the case with terms that retrospectively colonise prior practices. The applicability of digital twins has expanded dramatically since then. In manufacturing, for instance, digital twins underpin Industry 4.0 and its successor paradigms, enabling predictive maintenance, lifecycle optimisation, and human–machine collaboration [8]; and in the construction sector, digital twins facilitate accurate simulation of design scenarios before physical implementation, reducing costs and enabling early detection of critical flaws. This evolution is most visible in large-scale projects. Urban applications, perhaps the most ambitious in scale and implication, leverage digital twins to model smart cities, integrating traffic, land use, energy consumption, and citizen feedback into coherent, semantically grounded infrastructures. A digital twin of the city of Zurich is being developed to support sustainable urban planning [4]; and the new Singapore’s city-scale twin integrates data from sensors, satellites, and drones for growth management and environmental monitoring [9]. Moreover, the European Commission’s Destination Earth initiative envisions a planetary-scale digital twin to simulate climate dynamics, incorporating not only physical systems but also social and behavioural data [10]. This human-in-the-loop paradigm reflects a growing recognition that digital twins are not neutral mirrors but value-laden constructs. They encode assumptions about relevance, causality, and intervention and thus require governance frameworks that ensure transparency, ethical integrity, and stakeholder accountability [11].
When transposed to the cultural heritage domain, however, digital twins demand a more nuanced methodological framework. Unlike industrial or urban applications, where twins are created before or alongside their physical counterparts, heritage assets typically predate their digital representations by centuries and embody multiple temporal, material, and cultural layers. Consequently, their digital twins must account for both tangible and intangible dimensions, incorporating expressions of human creativity, identity, collective traditions, performances, and community practices that carry historical, artistic, social, and symbolic significance [12,13,14]. Within this perspective, digital twin technology has increasingly been tested across diverse areas of cultural heritage, producing significant though still heterogeneous results. In archaeology, digital twins have been used to document and virtually reconstruct sites and artefacts, providing new opportunities for exploring contexts that are otherwise fragmented or inaccessible [15]. The museum sector has adopted similar strategies for collections management and preventive conservation [16], with major institutions such as the Victoria & Albert Museum and the Louvre experimenting with large-scale digitisation to safeguard fragile artefacts and to expand access to audiences worldwide [17,18]. Monuments and architectural heritage have also been at the forefront of experimentation: the digital twin of Notre-Dame Cathedral, for instance, now functions as both a record of the building before the devastating fire of 2019 and a central resource for its ongoing restoration [19]. Comparable approaches have been applied to historic buildings, where parametric twins integrate sensor data to monitor indoor climate conditions and support conservation planning [20], as demonstrated by the cases of Löfstad Castle in Sweden [21] and Sant’Andrea’s church in Pistoia, Italy [14].
At the scale of urban and landscape heritage, projects such as the Great Wall Resource Management Information System [22] have generated detailed digital archives of UNESCO World Heritage sites across continents, demonstrating how digital twin technologies can be applied to cultural landscapes of vast territorial extension. Finally, in the field of restoration, digital twins enriched with AI-powered inpainting and 3D reconstruction have enabled the simulation of conservation treatments and the digital recovery of degraded artefacts, offering conservators a dynamic decision-support tool before interventions on the original object.
These examples demonstrate that, in most cases, applications of digital twin technology in cultural heritage remain predominantly tied to 3D visualisations, echoing late twentieth-century efforts to virtually reconstruct monuments and artefacts. As the model is evolving, the digital twin is gradually ceasing to be merely a static reflection, becoming a conceptual framework open to reinterpretation of operational and behavioural dimensions across disciplinary borders.

2.1. Digital Twins and Artificial Intelligence

Artificial Intelligence has increasingly emerged as a transformative force in the cultural heritage domain, offering powerful methods for analysing, interpreting, and even recreating the past. Its capacity to process vast datasets, recognise patterns, and generate new content has enriched the arsenal of digital tools available to scholars, conservators, and institutions in unprecedented ways. Applications span an impressively wide spectrum. Generative systems have been used to complete unfinished compositions by legendary composers, providing novel insights into artistic style and creative processes [23]. In parallel, advances in natural language processing have enabled the reconstruction and interpretation of ancient texts written in extinct or poorly understood languages, contributing to philology, epigraphy, and the study of lost literatures [24]. In archaeology and palaeoenvironmental studies, AI-driven reconstructions have recreated the aspect of long-lost environments and past landscapes, allowing researchers to simulate ecological conditions, visualise historical settings, and contextualise human activity [25,26].
Conservation and restoration represent another key area where AI has proven invaluable. In art history, recent advances have explored the innovative application of AI in the reconstruction and preservation of historical paintings, sculptures, ceramics, and other artefacts, offering new solutions to the challenges of traditional restoration [27]. The same technologies have been applied to questions of provenance, helping to trace the circulation of cultural objects across time and space through automated analysis of stylistic or material features [28]. Systems capable of detecting deterioration patterns in artworks and monuments provide guidance on preventive conservation strategies [29,30], while generative and inpainting algorithms have been successfully employed to support the virtual restoration of damaged paintings, sculptures, ceramics, and manuscripts [31]. These developments not only facilitate better-informed interventions but also open the possibility of testing conservation hypotheses in digital environments before their implementation on the original artefacts.
The incorporation of Artificial Intelligence into digital twin technologies is ushering in a significant conceptual and operational shift. No longer confined to functioning as elaborate digital counterparts of physical entities, digital twins are now evolving into systems endowed with interpretive, adaptive, and semantically structured intelligence, capable not just of replicating but also of understanding and responding to the complexities of their real-world counterparts. This transformation marks a notable epistemological threshold, recasting the digital twin as an active agent of interpretation rather than a passive conduit of representation. It is AI that propels this transition, furnishing digital twins with mechanisms of learning, inference, and autonomous decision-making [8]. As such, digital twins are no longer mere digital projections but are emerging as computational systems capable of analysis, simulation, and strategic recommendation [32]. Evidence of this transformation is manifesting across a broad range of domains. In manufacturing, for example, AI-enhanced digital twins support predictive maintenance, improve workflow design, and enable real-time responsiveness by interpreting sensory inputs and operational histories. Factories thus begin to resemble adaptive environments, governed not solely by pre-programmed rules but by evolving patterns of interaction and feedback [33]. In urban systems, the convergence of AI and digital twins is transforming how cities are modelled, understood, and governed. These intelligent frameworks are being applied to simulate environmental and infrastructural conditions—ranging from carbon emissions and heat distribution to traffic and pedestrian behaviour. As a result, municipalities are empowered to enact more nuanced and anticipatory interventions [34,35].
Importantly, the symbiosis between AI and digital twins is extending its reach beyond industrial or scientific contexts, entering realms such as cultural heritage with promising, often experimental, applications. One particularly innovative avenue involves the reconstruction of lost or undocumented heritage through AI-based image generation. Here, AI is employed to synthesise visualisations derived from oral testimonies and collective memory, complementing tangible-based HBIM methods [36].
In heritage-related scenarios more broadly, AI integration into digital twins is already proving beneficial for documentation, conservation, and public access. AI-driven twins are being employed to improve crisis management by integrating multi-sensor data and supporting real-time decision-making, with clear implications for the protection of heritage assets in emergency situations [37]. In cultural tourism contexts, the combination of digital twins and AI enables predictive monitoring, preventive conservation, and immersive storytelling, enhancing both preservation and visitor engagement [38]. At the same time, AI-enabled digital twins are increasingly able to reshape the relationship with artworks and heritage environments by allowing replication, remote access, and large-scale data integration [17]. More advanced approaches are exploring semantically grounded twins enhanced with memory-enabled AI agents, capable of reasoning over time and providing transparent, value-aligned support for long-term preservation planning [39]. Automation is expediting data collection, facilitating large-scale digitization, and enhancing accessibility via automated transcription and translation tools. At the same time, AI is processing real-time sensor data to detect threats and recommend countermeasures, increasing the system’s responsiveness to environmental risks or structural degradation. In this way, conservation scenarios can be virtually simulated, allowing interventions to be evaluated without exposing heritage objects to physical risk [14].
Despite these advances, the implementation of AI within digital twin ecosystems remains uneven and often constrained. Many applications persist in isolated pilot stages, lacking the infrastructural robustness or institutional momentum to scale effectively [40]. The interdependence of software, hardware, and human expertise highlights the need for deployment strategies that go beyond technical engineering to embrace socio-technical coordination [41]. Also, on a conceptual level, the integration of AI into the epistemic fabric of digital twins remains underdeveloped. While machine learning models excel in identifying correlations and predicting outcomes, they often fall short in contributing to the construction of coherent knowledge systems [42]. Nevertheless, progress continues steadily, and the expectation is that AI will increasingly endow digital twins with the capacity to fulfil their epistemic and operational potential. Yet this trajectory raises critical concerns about transparency and information quality. The internal logic of many AI systems remains opaque, making it difficult to trace how conclusions are reached—especially problematic in domains requiring interpretive subtlety, such as cultural heritage, urban planning, or medicine. Here, data in isolation is insufficient; meaningful engagement demands systems that can not only reason but also explain. An overreliance on autonomous inference without appropriate mechanisms for interpretability risks introducing errors, distortions, or even ethical blind spots. This situation prompts a foundational question: how far can epistemic authority be entrusted to systems incapable of articulating their own reasoning? Especially now that decisions once reserved for human deliberation start to be automated by systems that remain inaccessible to scrutiny, ensuring intelligibility and accountability is thus not a secondary concern but a prerequisite for ethical and sustainable implementation [11].

2.2. Ontologies: Illuminating Artificial Intelligence

In response to these fundamental questions, emerging approaches are foregrounding the importance of structured semantic architectures as a prerequisite for meaningful AI deployment. Rather than relying solely on data quantity or statistical modelling, these strategies emphasise the need for carefully curated knowledge frameworks, defining concepts, relationships, and ontological commitments, as foundational to advanced reasoning and explanation [43]. Ontologies, in this regard, are serving a dual role: they are enabling more nuanced and context-sensitive AI operations and also acting as safeguards against the risks of misinterpretation or hallucination [44]. Additionally, ontologies are employed in the development of explainable machine learning pipelines, addressing crucial aspects like feature categorisation and metadata description, which are fundamental for the transparency of the machine learning process [45]. By embedding formal logic and expert-curated structure within digital twin systems, ontologies help ensure that AI outputs remain anchored in coherent and verifiable knowledge domains. This allows intelligent systems not merely to function but to explain, interrogate, and evolve. In doing so, they contribute to the transformation of digital twins into epistemic infrastructures: semantic, reflexive, and dynamically coupled with the evolving landscape of human understanding [46]. This role of ontologies as a foundation for trustworthy AI is a crucial aspect, since they are able to provide more structured and explicit knowledge that enables AI systems to better articulate their reasoning. This extends beyond mere performance, addressing ethical concerns such as bias mitigation and transparency enhancement.

2.3. Towards the Cognitive Heritage Digital Twins

The Heritage Digital Twin currently being designed and developed [12,13,14,47] is shaped as a point of confluence; a space where these diverse technologies do not merely coexist but actively intertwine. It is precisely at this intersection that ontologies, Artificial Intelligence, and semantic infrastructures converge to generate a new kind of cultural intelligence. As evidenced in Section 2.1, typical applications of digital twins in heritage contexts largely focuse on high-resolution 3D models, often derived from laser scanning or photogrammetry. These visual representations, while technically sophisticated, are semantically shallow and remain inert unless complemented by rich contextual metadata and interpretive frameworks, offering a simulacrum in the visual sense but lacking interpretive depth.
Rather than reducing cultural heritage to its visual or geometric likeness, the paradigm of Heritage Digital Twin embraces the full constellation of digital documentation that relates to real-world cultural entities, whether movable, like artworks and artefacts; immovable, like monuments or architectural sites; or even intangible, such as rituals, traditions, and other expressions of cultural memory. This paradigm shifts the focus away from purely 3D-centred representations, proposing instead a holistic digital ecosystem that brings together visual depictions, textual accounts, scientific analyses, conservation records, and historical interpretations. It is a vision grounded in the belief that meaning arises not from isolated forms but from the interplay of all the fragments through which heritage is remembered, transmitted, studied, and understood.
Cultural entities, in fact, are not governed solely by physics but essentially by memory and cannot be fully described through geometry or physics alone since they carry meanings, contexts, and symbolisms that resist reduction to quantitative attributes. As a result, the replication of a heritage entity does not have to rest on form alone but requires a cognitive and semantic infrastructure and the ability to articulate not only what an object is but what it means, represents, and has undergone across time. As such, the digital twin in this domain demands not only a new technical infrastructure but a rethinking of its very ontological assumptions.
At the heart of any meaningful digital representation of cultural heritage lies a conceptual operation as much as a technical one: the deliberate choice to model reality through shared ontological commitments. Cultural entities are not inert objects, they are constructed through layers of description, interpretation and transmission. If their digital counterparts are to achieve more than a superficial resemblance, they must be capable of reflecting this diachronic and interpretive richness. Ontologies provide the necessary framework for this task: a rigorous grammar through which cultural knowledge may be structured, preserved, and subjected to critical inquiry. By expressing information as entities and relationships, ontologies enable a deeper representation of the cultural object’s life history, its transformations, and the meanings it acquires across time. In this framework, Artificial Intelligence can also reveal its full potential as a catalytic instrument capable of navigating, extracting, and reassembling meaning from the vast, fragmented archives of human knowledge and at the same time ingesting and digesting this knowledge to enhance its capabilities and serve as a privileged tool for exploring and interrogating the information embedded in the digital twin.
The culmination of this trajectory is thus the development of a new construct that unites the expressive power of ontologies with the interpretive capabilities of Artificial Intelligence in order to represent, reason about, and interact with cultural entities. Far from being a mere digital copy of its shape, the Heritage Digital Twin becomes an evolving semantic organism, an animated simulacrum (like Helen’s eidolon) which not only has its external resemblance but reproduces the cultural object in its entirety. In its most advanced form, the Heritage Digital Twin can become an agent (i.e., a cognitive entity) capable of dialogue, inference, and response within a meaningful digital continuum.
In this perspective, the present study proposes some innovative methodological aspects, building upon and extending existing practices. A distinctive contribution lies in the systematic combination of artificial intelligence and ontologies for the construction of Heritage Digital Twins (HDTs). While prior work in cultural heritage has often approached these technologies in isolation—relying either on AI without robust semantic guarantees or on ontological models lacking operational integration, as illustrated above—the proposed framework aims at unifying these two dimensions. Specifically, widely validated ontologies in cultural heritage documentation serve as the semantic backbone, ensuring the transparency, coherence, and interoperability of the data model. Concurrently, AI operates as the cognitive engine, dynamically extracting, enriching, and interpreting knowledge from the heterogeneous body of information that sustains the digital twin. This approach not only leverages the power of AI to process vast amounts of data—from analytical reports and curatorial records to conservation data—but also embeds this process within a structured, semantically rigorous framework. This ensures that the insights generated by the AI are verifiable and contextually relevant, a key advancement over previous approaches.
On this basis, this research pursues two main objectives: (i) to demonstrate how the proposed methodology facilitates the construction of Heritage Digital Twins by systematically examining and integrating heterogeneous documentation into a coherent and semantically grounded knowledge base and (ii) to show how this combined paradigm enhances the reactivity and cognitive potential of Heritage Digital Twins, enabling them to respond more effectively to analytical, conservation, and interpretative needs. The feasibility of these objectives is illustrated through two complementary case studies, aimed at illustrating the practical applicability of the proposed methodology: on one hand, ontologies and AI engage with the epistemic “archaeology” of cultural information, retrieving, structuring, and semantically articulating the dispersed and often latent knowledge embedded in visual and textual heritage records. On the other, they explore how this extracted knowledge, once formalised within a coherent ontological framework, can be reintroduced into AI systems to enhance their capacity for informed reasoning, contextual sensitivity, and traceable inference. A full account of the complete creation of a Cognitive Heritage Digital Twin would require a much larger study, given the scale and complexity of the task, encompassing vast amounts of documentation, data integration, and long-term monitoring. Far from exhausting the spectrum of possibilities opened by these technologies, this study exemplifies one of their most immediately impactful trajectories: the retrieval, organisation, and strategic deployment of knowledge embedded in cultural heritage documentation. The experiments conducted suggest that the procedures used could constitute a valuable contribution to anyone intending to address the same issues, especially in the cultural heritage sector where this type of experimentation is particularly needed.

3. Digital Twins and the Modelling of Knowledge

This paper builds upon the theoretical foundations introduced in our previous work on the Reactive Heritage Digital Twin and its associated ontology, offering a more operational perspective. It presents the design and implementation of AI-assisted semantic pipelines for the extraction, structuring, and reasoning over cultural information. By combining ontologies with intelligent processing components, it explores how to construct digital twins that are not only informative but also sentient in their logic and responsive in their behaviour. More specifically, this paper examines two complementary uses of Artificial Intelligence within this framework, situated at opposite ends of the semantic pipeline. On one side, AI is deployed for the extraction of knowledge from unstructured textual sources, in particular, heritage science reports, and its subsequent alignment with ontological structures, resulting in the enrichment of the digital twin’s semantic graph with scientific information. On the other hand, the populated graph itself becomes a resource for training AI systems, enabling them to perform tasks such as querying semantic information through natural language or assisting in the extraction of new content. This virtuous cycle of knowledge generation and reinforcement creates a recursive architecture in which semantic data and intelligent agents co-evolve, gradually enhancing the digital twin’s cognitive capacity. The goal is to show how this technological convergence can support the development of truly intelligent Heritage Digital Twins: systems that do not merely reflect culture but actively participate in its documentation, understanding, and preservation.
The conceptual architecture upon which the entire system is constructed is the Reactive Heritage Digital Twin Ontology (RHDTO), an extension blossomed from the solid foundations of the CIDOC CRM [48] as a means to organise, interconnect, and semantically sustain the heterogeneous data that forms the informational substrate of cultural digital twins. Among the various ontological frameworks developed for the heritage domain, the CIDOC Conceptual Reference Model (CRM) is an internationally recognised standard (ISO 21127:2023 [49]). More than a model, CIDOC CRM constitutes an ecosystem that includes domain-specific extensions, such as CRMsci for scientific observation [50], CRMdig for digital provenance [51], and CRMhs for heritage science [52], that offer a nuanced vocabulary capable of accommodating the multifaceted nature of cultural data. The adoption of CIDOC CRM offers a new path forward, disclosing the possibility to describe not just the form of an object but the historical events, actors, techniques, and materials that constitute its biography. It is this ontological fluency, this capacity to speak the shared language of heritage knowledge, that renders the CIDOC CRM ecosystem a cornerstone of semantic continuity across the heritage information landscape.
In some previous works, the structural components of the RHDTO have been articulated, delineating its main dimensions: the documentary one, which captures the breadth of scientific, historical, and multimedia documentation [12,13]; the reactive one, which models the operational mechanisms for real-world interaction through sensors, deciders, and actuators [14]; and the intelligent one, in which Artificial Intelligence is integrated to interpret, enrich, and activate knowledge [53]. To operationalise these dimensions, the RHDTO introduces a set of classes and properties designed to model the multifaceted nature of heritage entities and their digital counterparts with semantic precision. At the core lies the class HC1 Heritage Entity, a general abstraction encompassing both tangible (HC3 Tangible Aspect) and intangible (HC4 Intangible Aspect) dimensions of cultural heritage. The digital twin itself is represented by HC2 Heritage Digital Twin, conceived as a structured network of informational components, including digital reproductions, documentation, and the activities underpinning their creation and management. Notably, Ref. [53] illustrates how AI can serve as the cognitive engine of the digital twin, supporting analytical reasoning and decision-making within key components such as the Decider and the Actuator, thereby enhancing the system’s ability to respond to real-world conditions in a meaningful and context-aware manner. Each of these facets are modelled by means of dedicated classes, such as HC10 Decider and HC15 AI Component, and contribute to the evolving ecology of digital twins. Building on this, the subsequent sections describe several ways in which AI can be practically used for this purpose, offering an overview of the most recent methodologies and technologies available.

4. From Fragments to Knowledge

4.1. Natural Language Processing of Textual Documentation

NLP focuses on enabling machines to understand, interpret, and generate human language. As simple as this may sound from an agnostic point of view, handling human language is one of the most difficult tasks a machine can undertake. Unlike programming languages, which are based on mathematical rules and are explicitly designed to be machine-understandable, natural language is not as easily interpreted or represented in numerical form.
Over the past decade, the field of NLP has undergone a major transformation with the advent of deep learning and, more recently, Large Language Models (LLMs). These models are based on the Transformer architecture, which introduced the so-called attention mechanism:
“Given a set of vector values and a vector query, attention is a technique to compute a weighted sum of the values, dependent on the query.” [54]
In particular, LLMs use a specialised version called self-attention.
In traditional attention mechanisms, each query maps exactly to one key–value pair, while in self-attention each query matches each key to varying degrees and then the returned result is a sum of values weighted by the query–key match. This mechanism allows the model to assign importance to different tokens (which are often single words) within a sequence, enabling it to focus on the most relevant parts of the input when generating output.
In addition to this, LLMs are trained on vast, multilingual corpora [55] and have demonstrated impressive capabilities across tasks [56] such as machine translation, question answering, summarisation, and semantic search. Their ability to perform complex tasks such as transfer learning, a machine learning technique in which the knowledge learned from a task is re-used to boost performance on a related task, few-shot generalisation [57], and contextual language modelling [58] has opened new frontiers in computational language understanding, particularly in complex or previously underexplored text domains [59].
One such domain is cultural heritage [60], where textual artefacts may include ancient manuscripts, epigraphic inscriptions, archival inventories, scanned PDFs of scientific analyses, and handwritten letters [61]. These materials offer rich but highly challenging content for computational analysis, often exhibiting historical language variants, inconsistent orthography, multilingual passages, and physical degradation, all of which complicate standard NLP workflows.
Recent efforts by computer scientists have focused on adapting LLMs and developing domain-specific NLP pipelines to address these challenges. Techniques such as fine-tuning on specialised corpora, integrating Optical Character Recognition (OCR) and Handwritten Text Recognition (HTR) outputs [62], and combining textual data with metadata or image features [63] are increasingly used to support tasks like:
  • named entity recognition
  • semantic enrichment
  • machine-assisted transcription
  • cross-document linking
As a result, state-of-the-art NLP is becoming an essential component in the digital preservation, accessibility, and scholarly analysis of cultural heritage texts.

4.2. LLMs for Identification and Semantic Extraction of Relevant Entities and Relationships

A critical challenge in the digital processing of cultural heritage texts lies in the identification and extraction of entities and relationships, a foundational step for building structured knowledge, supporting semantic search, and enabling automated reasoning.
Traditionally, this task has been addressed through Named Entity Recognition (NER) pipelines, which label predefined categories such as “person”, “location”, “organisation”, or “date” in unstructured text. Over the last decade, state-of-the-art NER systems have been greatly enhanced by transformer-based models like BERT and RoBERTa, showing high accuracy on well-structured, contemporary texts. However, these systems often struggle when applied to historical documents [64], multilingual corpora, or domain-specific ontologies, where named entities are ambiguous, infrequent, incompletely annotated, or entangled in complex relational structures (even though there are some successfully trained version of some specific domains, such as ArchaeoBERT [65]). In the context of cultural heritage, standard NER often proves insufficient. Texts, such as archival inventories, epigraphs, marginalia, or curatorial notes, frequently reference entities in indirect, partial, or obsolete forms [66]. Furthermore, the relevant concepts often extend beyond standard categories to include historical events, material properties, artistic styles, ritual roles, or geopolitical shifts, many of which are not well covered by off-the-shelf NER models [67].
These limitations pose significant barriers to aligning extracted information with structured ontologies or linking it to a Knowledge Graph.
Recent research has increasingly highlighted the intersection of LLMs and KGs, pointing to a paradigm shift in knowledge representation. On one side, LLMs are used to augment KGs by supporting knowledge extraction, construction, and refinement. On the other, KGs are used to augment LLMs in tasks such as training, prompt learning, or knowledge grounding [68,69]. These works underline a move towards hybrid models and pipelines that integrate both explicit ontological knowledge and the parametric knowledge encoded within LLMs.
KG construction itself is a complex process that requires the integration of information both from structured and unstructured data. Traditional pipelines treat these sources in isolation, struggling with the heterogeneity and noise typical of cultural heritage data. LLMs, by contrast, are trained across diverse sources and exhibit strong performance on tasks such as knowledge extraction [70], entity resolution [71], and relation alignment.
An LLM-based approach is employed in this experiment, enabling flexible, context-aware entity and relation extraction in the cultural heritage domain, eliminating reliance on pre-trained NER schemas.
Instead of classifying entities into fixed types, the LLM is prompted to identify relevant semantic units and their relationships in natural language, guided by domain-specific ontology definitions. This enables extraction and alignment of non-standard or composite entities and their mapping to structured entities in a customised Cultural Heritage Knowledge Graph.
This prompt-based methodology further supports the extraction of complex relationships, including part-whole, temporal, provenance, and scientific analysis relations, which are essential for modelling cultural heritage information in ontological formats.
By bypassing traditional NER in favour of in-context reasoning with LLMs, multiple tasks, such as entity disambiguation, coreference resolution, attribute assignment, and relation typing, are integrated into a single unified pipeline.
While LLM fine-tuning methods can leverage textual information from individual entities, formal semantics are often underutilised. Beyond template-based fine-tuning, additional LLM techniques, such as prompt learning and instruction tuning, present promising avenues to enrich semantic understanding [72,73]. Moreover, cultural heritage applications often rely on cross-domain knowledge, highlighting the importance of ontology alignment, so the identification of cross-ontology mappings among concepts, instances, and properties with equivalence, integration, or membership relationships can occur.
Traditional alignment systems, such as LogMap [74], rely heavily on lexical matching and symbolic reasoning, but this is often not enough. Exploiting textual meta-information through LLMs offers a promising strategy for ontology alignment: even though pretrained models like BERT have been fine-tuned for this task [75], state-of-the-art LLMs are yet to be explored in the context. Our experiments show that this LLM-driven strategy significantly improves recall and semantic accuracy in entity-linking tasks when compared to traditional NER-based pipelines, particularly when dealing with fragmented historical records or low-resource languages. Moreover, the use of an open-source, locally deployable model ensures data privacy, customisability, and scalability, meeting the needs of cultural institutions and digital humanities projects that require transparent, interpretable, and adaptable AI tools.
Transitioning extracted information into an ontological encoding makes it unambiguous, facilitating integration into the Cultural Heritage Knowledge Graph and subsequent incorporation into Heritage Digital Twins, thus supporting richer and more semantically coherent digital representations of cultural assets.

4.3. Building Semantic Knowledge Graphs from Extracted Information

The process of encoding NLP-extracted data to ontological structures is never a mere act of correspondence; it is a conceptual translation that requires both technical rigour and hermeneutic sensitivity. Information drawn from textual descriptions, archival records, scientific datasets, or visual analysis carries within it layers of context, intention, and uncertainty. To encode such data into an ontological framework is to interpret it through a lens that seeks formal clarity without erasing epistemic nuance.
One of the major strengths of the RHDTO model lies in its ability to function as a connective tissue between datasets that originate from heterogeneous contexts, a conceptual harmonisation that permits the assembly of coherent knowledge graphs without sacrificing domain specificity. RHDTO provides the necessary paraphernalia for such knowledge transformation and transcoding, turning textual evidence and visual data into computable knowledge while preserving their contextual richness. This is particularly relevant in domains like heritage science, where the use of these conceptual tools allows for a fine-grained articulation of experimental procedures and datasets, as will be illustrated in Section 6. This semantic articulation reshapes the cognitive architecture through which heritage information is organised, shared, and understood, transforming the ontological framework of the knowledge graph in a philosophical intermediary between language and logic, between human interpretation and machine understanding [76]. This way, the digital twin is endowed with a cognitive dimension. It no longer merely stores or displays data but interprets and internalises it within an ontological order that mirrors, albeit partially and provisionally, the complex web of human understanding.
Artificial Intelligence, particularly in its recent semantic incarnations, also plays an increasingly central role in this mapping process. AI systems can retrieve entities, classify relationships, and extract candidate assertions from unstructured sources and convert them in ontology-grounded semantic triples. This is one of the cases in which the synergy between AI and ontologies reveals its full potential, the former accelerating the identification of candidate knowledge, the latter ensuring that such knowledge is meaningfully situated within a coherent and reusable conceptual space. Yet the true work of semantic alignment remains interpretive, demanding expert judgement in choosing the appropriate classes, properties, and event structures through which to articulate extracted content within the framework of the knowledge graph.

5. From Knowledge to Intelligence

5.1. Using Knowledge Graphs to Build Knowledge-Enriched AI Agents

Once knowledge crystallises in the form of semantic graphs, it can be used to nurture and enhance AIs. The complexity of cultural heritage data demands the use of intelligent systems that can reason over both explicit content and the nuanced context behind it. Knowledge graphs, for this purpose, have emerged as a powerful tool to build knowledge-enriched Artificial Intelligence systems. Information is structured to preserve its rich semantics and remains machine-interpretable, which enables AI systems to achieve context-aware understanding and inference instead of limiting themselves to surface-level pattern recognition [77,78].
Cultural heritage tells a story, and as stories are composed of interrelated entities, they must be represented in such fashion. A knowledge base with its nodes and edges perfectly captures this web of elements, grounded in formal ontologies [78,79], and CIDOC CRM and its extensions play a crucial role in offering a shared conceptual schema that aligns historical data with contemporary semantic technologies [78,80,81].
In recent years, there has been a growing emphasis on reinforcing the reasoning capabilities of LLMs by integrating them with external knowledge structures, particularly knowledge graphs. While LLMs such as GPT, BERT, and Mistral [82,83,84] have demonstrated remarkable abilities in generating fluent and plausible language, they remain prone to hallucinations, generating incorrect or misleading content especially in high-stakes domains like healthcare or science [85].
These failures often stem from limitations in their training objective, typically maximising the log-likelihood of the next token without robust grounding in factual correctness.
To address this, researchers are increasingly integrating LLMs with knowledge graphs, which encode structured, semantically rich, and traceable factual knowledge [86].
Unlike opaque neural representations, knowledge graphs enable models to reason over explicit entities and relations, align responses to verifiable facts, and offer interpretability via provenance metadata [87,88].
This integration allows models not only to retrieve precise domain-specific information but also to construct explainable inference chains, thus improving transparency and trust. Moreover, the inclusion of knowledge graphs supports granular and contextually relevant knowledge retrieval, outperforming naive augmentation with random or noisy data [89]. By aligning entities and relationships from LLM outputs with ontologies and knowledge bases, systems can avoid shallow text pattern recognition and instead achieve conceptual generalisation and symbolic reasoning.
This approach also opens the door to continuous learning pipelines, where AI outputs are validated against, or enriched by, structured data and subsequently reintroduced into the model in the form of feedback loops. These loops allow AI systems to iteratively refine their knowledge representations, adjust for emerging information, and correct previously observed errors. Combined with recent advances in multimodal and temporal input integration—e.g., linking structured data with images or spatial metadata [68]—the use of knowledge graphs is becoming central to the development of knowledge-enriched, context-aware AI systems that are both accurate and adaptable. As structured data continues to permeate all domains of life, designing robust mechanisms to represent and integrate this data into LLMs remains a crucial and active research frontier.

5.2. Embedding Cultural Heritage Knowledge Graphs

To date, much of the literature has concentrated on the construction of knowledge graphs, focusing on entity recognition, ontology alignment, and data integration with ontologies such as CIDOC CRM [90,91]. More recently, however, researchers have begun to investigate the potential of graph representation learning as a means of exploiting these graphs for retrieval, enrichment, and semantic discovery.
A notable example is the CIDOC2VEC approach [92], which develops embeddings from CIDOC-CRM-compliant graphs. The method employs relative sentence walks through the KG to generate sequences of entities, which are then embedded using standard distributional techniques from natural language processing. These embeddings enable similarity-based retrieval and recommendation of cultural heritage objects across collections. CIDOC2VEC represents a significant advancement because it demonstrates how embeddings can facilitate semantic interoperability across institutions, allowing users to discover related artefacts even in the absence of explicitly asserted links.
Nevertheless, a critical limitation of CIDOC2VEC is its entity-centered, relation-agnostic design. By treating walks as linear sequences of entities, the approach largely ignores the semantic role of relations defined in CIDOC CRM, such as the distinction between P108 has produced (linking production events to artefacts) and P50 has current keeper (linking artefacts to institutions). While this simplification improves scalability, it results in embeddings that capture contextual co-occurrence but lack sensitivity to the multi-relational structure that is central to CRM’s expressiveness. Consequently, retrieval is biased toward entity similarity rather than relation-aware reasoning, limiting its ability to support complex queries (e.g., retrieving artefacts created in the same workshop but preserved in different sites).
In contrast, recent advances in graph neural networks (GNNs), and particularly composition-based graph convolutional models such as CompGCN [93], explicitly model both entities and relations during the embedding process. By applying relation-specific transformations during message passing, CompGCN embeddings encode not only which entities are related but also how they are related, producing richer and more semantically faithful representations. In mainstream KG domains (e.g., Freebase, DBpedia), this relation-aware modelling has been shown to significantly improve link prediction and semantic querying [93,94].
Yet, from the review carried out in this research work, relation-aware graph convolutional models have not been applied in the querying phase of cultural heritage KGs, particularly those structured with CIDOC CRM. Existing approaches remain entity-focused, prioritizing similarity-based retrieval (as in CIDOC2VEC) or classification tasks (e.g., GCNBoost [95]) rather than supporting relation-aware querying and ranking. This presents a clear research gap.
This study addresses said gap by exploring the use of CompGCN embeddings to enhance querying in the RHDTO grounded knowledge graph. Rather than focusing exclusively on graph construction or classification, a proposal is extended to apply embeddings post-construction in the querying phase, enabling relation-sensitive similarity scoring across entities and events. This contributes a novel paradigm for semantic search and exploratory browsing, bridging the rigour of ontology-driven reasoning with the flexibility of embedding-based retrieval.

5.3. Access to Heritage Data in Natural Language

As the scale, scope, and complexity of data increase, the need to simplify its querying methods does so equally. Past a time where only select individuals could access certain information thanks to their expertise in the topic and research skills, information is now at everyone’s reach, a half-thought and a click away. Not only archaeologists, scientists, or other specialists may wish to browse the knowledge base, so in the context of Heritage Digital Twins, traditional querying methods would rather prove obsolete in the face of approaches utilising natural language as a medium [96].
If a user expresses their inquiry in their own language, the translation layer—natural language understanding (NLU)—bridges the gap between human intention and structured semantic content [97]. The query gains a certain level of interpretability, relaxing constraints and allowing for fuzzy reasoning during retrieval across both structural and semantic levels [98]. Not only does this present an advantage in query guarantee (retrieving at least one result) but also provides users enough feedback to adjust and shape their queries and best attract the answers they seek within the intricate knowledge graph.
At the core of this capability is a combination of tasks including semantic parsing, entity linking, intent classification, and relationship inference to decompose and interpret the natural language queries [99]. Thanks to recent advances in prompt-based large language models, using separately trained components for each of these tasks is no longer a requirement, as LLMs offer an integrated, zero, or few-shot approach to parsing natural language [100]. In the context of cultural heritage, where user queries are often imprecise, explanatory, and laden with historical, geographical, or culturally specific terms, properly instructed LLMs can produce a variety of structured outputs [77]. The LLM prompt construction includes the general context of the query to focus the interpretation, indications on the what and how of information extraction, and any additional enrichment LLMs should apply to their output. As such, the latter can be in the form of data triples—the building blocks of a graph structure [98].
Intuitively, these triples would be converted into SPARQL fragments to filter the knowledge graph data by mentioned entities, and the results would be layered according to the specific query relationships. The SPARQL query, then, may only be constructed after correct assignment of ontology labels and properties to the processed query elements—a task which would either require the involvement of an expert for labelling and validation or the implementation of an additional layer of entity and relationship classification using an ontology-trained AI model [101]. However, the usage of multiple hierarchical ontologies, one more complex than the other, renders this task more complicated than not, and while the latter option is more viable, the rigidity of SPARQL querying remains a risk; in the case of misclassification, the query will yield no results although they might exist [102].
To address this challenge, a hybrid query execution is adopted that bypasses strict SPARQL-based querying and instead leverages graph neural networks and knowledge embeddings, allowing the system to execute natural language queries in a semantically flexible and structurally tolerant way, specifically via Composition-based Graph Convolutional Networks (CompGCNs).
Following the NLU stage, the nodes represent candidate entities, types, or conceptual elements (e.g., “cathedral”, “bronze”, “Charles II”), and edges represent inferred relations or semantic dependencies (e.g., “located_in”, “issued_by”, “made_of”). The constructed query graph represents the user’s intent directly, without prematurely enforcing alignment to any specific ontology schema, while capturing its semantic and relational structure.
For Heritage Digital Twins, similarity extends beyond the simple attribute matching but rather enters the world of semantic, structural, and contextual equivalence. It is rare—if at all possible—for artefacts, historical and archaeological remains, or intangible heritage to be identical. Instead, they show overlap in historical context or provenance, form, function or symbolism [103], all types of data that are hosted and structured in the knowledge graph according to the RHDTO.
For tangible heritage, this work investigates similarity computation at the following levels, for example, to identify:
  • pieces that belong to the same object,
  • different objects that were excavated at the same site but from different locations,
  • different objects that were crafted in the same workshop,
  • different objects made of material originating from the same place; inferred from the data of scientific analysis reports,
  • different objects made by the same artist.
This demonstrates the great advantage of similarity computation in querying, since thus far in existing works, most of these results would not be retrieved by a basic querying system alone. In fact, some of these objectives require reliance on visual input [104] such as images or 3D models, but by using intelligent AI models, it may be possible in most cases to identify such similarities just based on metadata stored within the knowledge graph [105].

6. Case Study: Enriching Digital Twins with Heritage Science Data

As a testbed for exploring the integration and synergy between AI techniques and ontological modelling, some reports generated by analysis activities in the heritage science field were selected, part of the documentation of the ARTEMIS initiative [106].

6.1. Application of AI-Driven Semantic Pipelines

A modular AI-based pipeline was designed to extract structured semantic knowledge from unstructured scientific texts belonging to the cultural heritage domain, with a focus on modern preservation and restoration techniques such as, but not limited to, X-Ray Fluorescence (XRF) and 3D modelling techniques.
The system is particularly tailored to support the enrichment and population of an ontology-driven knowledge graph designed to be an efficient base for the construction of Heritage Digital Twins, but its architecture is flexible enough to support other knowledge-driven applications.
The defined methodology combines several state-of-the-art open-source technologies, including Natural Language Processing, Large Language Models, semantic similarity search, and ontology-based reasoning.
Textual data was primarily sourced from the ARTEMIS initiative documentation, consisting of complex PDF files with multi-column layouts, embedded images, tables, and special characters. The preprocessing and chunking pipeline consisted of the following steps:
  • PDF parsing and image extraction: text and images extraction has been performed with PyMuPDF [107], a Python (Python 3.12.7, PyMuPDF 1.25.5) library able to handle even complex textual layout. Images from PDFs were also extracted and kept separately for further experiments.
  • Text preprocessing: SpaCy [108] library has been used to perform tokenisation, sentence segmentation, and syntactic parsing to reduce noise and generally improve the handling of textual data by machines.
  • Sentence-aware chunking: to enable efficient processing by Transformer-based LLMs, the texts were then segmented into smaller, semantically coherent units called “chunks”. Sentences were aggregated sequentially until the token limit ( in this experiment the limit was 300 token) was reached to avoid splitting syntactically or semantically dependent clauses. This size has been chosen to balance the model’s context window constraints with our need to have meaningful textual context. This segmentation allows the model to focus on localised contexts, ensuring that the extracted knowledge is specific and grounded in the source material.
  • Text normalisation: lowercased all text, removed punctuation, standardised whitespace, and ensured consistent character encoding for compatibility with embedding models.
Once the text chunks were generated and normalised, the pipeline employed a combination of embedding-based semantic analysis and instruction-tuned language modelling to perform ontology term extraction and knowledge graph population. Candidate ontology terms extracted from each chunk were first encoded into dense vector representations using the SentenceTransformer model all-MiniLM-L6-v2 [109]. A pairwise cosine similarity matrix was then computed across all term embeddings to identify semantically similar terms. Terms which exceeded a chosen similarity threshold of 0.85 were clustered together, effectively collapsing synonyms and variant spellings into unified conceptual entities. The pairwise cosine similarity matrix is given by:
S = D 1 X X D 1 ,
where
D = diag x 1 , x 2 , , x n .
This semantic deduplication approach allowed for a more robust normalization than conventional string-matching methods, ensuring terminological consistency across the corpus.
Term extraction was then performed using Mistral 7B Instruct [110], an instruction-tuned LLM selected for its strong performance and open-source transparency, which guarantees the reproducibility of its outputs.
The prompting strategy was explicitly designed to extract only the entities and concepts relevant to the cultural heritage domain.
Each chunk of text was provided to the model with instructions to identify modern scientific analysis techniques, materials, instruments, and experimental procedures while excluding any narrative or explanatory content.
Prompts were crafted to elicit concise, list-form outputs, containing only the terms of interest. By carefully structuring these prompts, the model could perform zero-shot extraction of domain-specific ontology terms without relying on supervised training data, ensuring that each output was both focused and reproducible.
To ensure the reliability of the extracted terms, a human-in-the-loop validation step was implemented, allowing domain experts to manually review and select relevant terms. This hybrid approach balanced automated efficiency with expert oversight, mitigating the risk of hallucinated or irrelevant outputs while maintaining the accuracy of the ontology term set.
A key innovation of this pipeline lies in the retrieval of semantic context for each validated term: while processes like Term Extraction or Named Entity Recognition successfully provide a list of relevant concepts, their meaning is often best understood in relation to the text from which they originate, and this is especially true in our case, where the data needs to be mapped following domain ontologies rules.
To capture this context, both the refined ontology terms and the original text chunks were embedded using the same SentenceTransformers model, and a FAISS similarity index was constructed over the chunk embeddings.
FAISS [111] is an open-source library for efficient similarity search and clustering of dense vectors, optimised for high-dimensional data, such as embeddings from neural networks. The way it works is by returning k nearest neighbors of a query vector, typically with their indices and distances/similarity scores.
With the relevant contextual chunks retrieved, Mistral 7B Instruct was employed a second time to generate RDF triples.
The model was prompted to extract concise semantic triples aligned both structurally and semantically with existing cultural heritage ontologies. This resulted in a set of triples that formally represents relationships between terms while preserving traceability to the source text.
The final outputs of this AI-assisted pipeline include:
  • a validated list of ontology-relevant terms, stored in JSON format,
  • a collection of RDF triples, structured and domain ontology-aligned,
  • a mapping between each term and its original context, ensuring interpretability and traceability.
Figure 1 is a diagram describing graphically the aforementioned pipeline:
The underlying conceptual model chosen for this effort is CRMhs [52], an ontological component of the CIDOC CRM ecosystem developed to represent the documentation and operational structure of scientific analyses carried out on cultural heritage objects. As an extension of the CIDOC CRM, CRMhs integrates seamlessly with RHDTO and gives it the ability to formalise the domain of heritage science by introducing specific classes and properties that describe scientific activities (such as HS3 Analysis), instruments (e.g., HS11 Scientific Instrument), datasets, sampling procedures, and their relationships to both the physical object under investigation and the broader research context. It supports the representation of tangible and intangible entities involved in the scientific process, such as samples, areas of interest, methods employed, and results obtained. The model also accounts for the configuration of complex devices, their software components, and even the parameters used during analysis. Moreover, since the ARTEMIS ontological architecture is intentionally modular, reflecting the diversity and complexity of the domains it seeks to represent, the project also employs the AO-Cat ontology [112] developed within the ARIADNE initiative [113] to describe datasets and their specific characteristics, including provenance, access conditions, and content typologies. Our encoding experiments, also using AO-Cat to encode these entities, have fully demonstrated how these models are able to interact perfectly with each other, allowing the construction of semantic graphs that, despite being internally heterogeneous, constantly maintain a solid semantic coherence.
A general overview of the ontological module designed by ARTEMIS to model scientific data is shown in Figure 2.
By anchoring AI outputs to a domain-specific ontology, the system ensures that extracted data is both semantically coherent and interoperable. This ontological framework thus provides the structure for an AI-assisted information extraction workflow aimed at transforming heritage science reports into semantically enriched data. Using the categories defined by this semantic ecosystem as guidance, natural language processing tools are applied to identify and annotate references to analytical techniques, measurements, instruments, and related entities. The extracted knowledge is then converted into structured triples, for instance, in Turtle format, encoded using the grammar provided by the ontological model employed and ingested in the knowledge graph.
The following examples show some results of the pipeline tested in [114] about the use of Macro X-ray Fluorescence Scanning on a Raffaello’s Portrait of Leo X. Starting from the original PDF, the JSON file containing the triples was successfully and automatically extracted, encoded in Turtle format, and ingested into the ARTEMIS knowledge base to visualise the graph. A fragment of the Turtle encoding (below) shows the ontological representation of Raffaello’s painting as a cultural object by means of the HC3 class of the RHDT Ontology.
artemis:LeoX_Portrait_Study_Object a rhdto:HC3_Tangible_Entity ;
  rdfs:label "Portrait of Leo X" ;
  owl:sameAs <http://www.wikidata.org/entity/Q5597> ;
  crm:P102_has_title "Portrait of Leo X with Cardinals" ;
  rhdto:HP11_was_made_by artemis:Artist_Raffaello ;
  crm:P45_consists_of "Oil on wood" ;
  rhdto:HP12_was_made_within artemis:LeoX_Portrait_Period ;
  crm:P55_has_current_location artemis:LeoX_Portrait_Place .
		
The following Turtle fragment, instead, shows the encoding of the analysis event conducted by scholars of the Italian Institute of Nuclear Physics (INFN) on the same cultural object, the way in which CRMhs properties are used to specify the instrumentation and methodology employed, and the location and dates on which the analyses were performed.
artemis:Analysis_MA_XRF_LeoX a crmhs:HS3_Analysis ;
  rdfs:label "Macro X-ray fluorescence analysis" ;
  crm:P2_has_type "MA-XRF" ;
  crmhs:HSP1_has_activity_title "MA-XRF analysis on Leo X Portrait" ;
  crm:P3_has_note "Macro X-ray fluorescence analysis" ;
  crmhs:HSP4_analysed artemis:LeoX_Portrait_Study_Object ;
  crmhs:HSP6_used_method artemis:XRF_Method ;
  crmhs:HSP8_used_instrument artemis:Device_INFN_CHNET_Scanner ;
  crmhs:HSP9_used_component artemis:Component_XRay_Tube_Moxtek ,
                    artemis:Component_SDD_Amptek ,
                    artemis:Component_Telemeter_Keyence ;
  crmhs:HSP10_used_software artemis:INFN-CHnet_Software ;
  crmhs:HSP13_used_settings artemis:Settings_001 ;
  crmhs:HSP16_produced_dataset artemis:XRF_LeoX_Resulting_Paper ;
  crm:P14_carried_out_by artemis:Group_INFN ;
  crm:P11_had_participant artemis:Person_Lorenzo_Giuntini ;
  crm:P4_has_time-span artemis:Analysis_Time_Frame ;
  crm:P7_took_place_at artemis:Analysis_Place .
		
A complete encoding of this entities, in Turtle format, can be found in the Supplementary File accompanying this paper.
These outputs can be ingested into Heritage Digital Twin Knowledge Graphs, enabling advanced features such as semantic querying, dynamic visualisation, and automated knowledge integration from heterogeneous scientific sources and the subsequent feeding of AIs for advanced data analysis and interactions.
A visual representation of a section of the Digital Twin Knowledge Graph enriched with semantic context through our AI-assisted process is shown in Figure 3 below.

6.2. Populating and Querying the ARTEMIS Knowledge Graph

The following subsections describe the second phase of the experiment pertaining to the querying of the knowledge graph, for which the steps are illustrated in the process diagram in Figure 4.
Once the HDT knowledge graph has been built and the cultural heritage data is semantically structured, the next step is implementing the querying system, which allows us to access it. This process involves a few steps, beginning with an additional preprocessing stage. This time, we are not preprocessing textual documents but rather preparing the stored data for artificial intelligence model training by indexing each entity and relation with a unique ID and engineering the graph data as described next. The ontology-based graph structure resulting from the preceding work maintains the information in a parametrised way, linking all attributes through property edges, including the node types (e.g., by means of the P2 has type property of CIDOC CRM). It is good practice for data archival, consistency, and reusability; however, it poses particular challenges for an AI model, as then the nodes do not encapsulate enough meaningful information to distinguish them from others. To address this, the knowledge graph is restructured so that each node in the graph corresponds to an instance of an RHDTO class. Therefore, the type of each entity is embedded in its node as a static feature, along with any human-readable label, title, or description it is linked to, for all of which simple embeddings were created using SentenceBERT. An example of the resulting node in JSON format is the following:
 { "id": "obj_0001",
  "label": "Mona Lisa",
  "type": "E22_Man-Made_Object",
  "features": {
  "label_embedding": [0.23, 0.78, ...],
  "type_embedding": [0.12, -0.09, ...] }
		
The next step of the process is to feed the new knowledge graph to an AI model and begin training. Though, grounding the knowledge graph in the rich RHTDO makes it too complex for standard neural networks to capture its nuanced relational structure of data. The selected model, CompGCN, explicitly models both nodes and relations jointly. For cultural heritage data where each connection has a semantic meaning that depends on the nature of the relationship, it is critical to be able to make a difference between—for example, whether a person created an object, participated in an event, or curated an exhibition [93]. CompGCN looks at each of these relations as its own learnable element and trains to learn their differences. It also accounts for the directionality of relationships; in the knowledge graph, many relationships are inherently directional and semantically loaded, as there is an asymmetry in reading them (e.g., “was influenced by” and “influenced”). The network learns this by identifying the roles of the relations and the flow of information between nodes [93].
Implementation-wise, the graph is transformed into triple format or adjacency lists, and each relation is modelled explicitly with its inverse counterpart. CompGCN then uses a message-passing mechanism that jointly embeds nodes and relations to spread information throughout the graph during training. The model uses composition operators on the embedding vectors to integrate relation semantics into the neighbourhood aggregation, which allows the network to model logic and semantics of the paths traversed additionally to their proximity. Node embeddings are updated at each graph convolution layer according to the unique semantic composition of the connecting relation and the characteristics of their neighbours [93].

6.3. Message Passing in CompGCN for Cultural Heritage Knowledge Graphs

In CompGCN, the message passing mechanism extends traditional GCNs by explicitly incorporating relation semantics [93,94]. For each triple ( h , r , t ) , the neighbour embedding t ( k ) is composed with the relation embedding r ( k ) using a composition operator ϕ , e.g.,
m h ( t , r ) ( k ) = ϕ ( t ( k ) , r ( k ) ) ,
where ϕ can be addition, multiplication, or circular correlation. Each entity representation is then updated by aggregating over its adjacency list:
h ( k + 1 ) = σ ( r , t ) Adj ( h ) W r ( k ) · ϕ ( t ( k ) , r ( k ) ) + W 0 ( k ) h ( k ) ,
where W r ( k ) are relation-specific weight matrices and σ is a non-linear activation. Unlike earlier relational GCNs, CompGCN also updates relation embeddings:
r ( k + 1 ) = W r ( k ) r ( k ) .
This joint update allows both entities and relations to evolve during training, which is particularly relevant in cultural heritage graphs where relations (was produced by, participated in, located at) carry rich semantics. For example, the embedding of an entity such as a Vase is updated not only from its neighbouring Potter (was produced by) but also from its participation in a Firing Event [115]. After three layers, entity and relation embeddings are evaluated with a TransE scoring function [116]:
f ( h , r , t ) = h + r t 2 ,
and optimised using a margin ranking loss to ensure valid triples score higher than corrupted ones. This allows embeddings to capture both contextual and relational meaning, enabling similarity-based querying of heritage data. The following schema in Figure 5 illustrates this aggregation.

6.4. Similarity Computation Between the Query and KG Embeddings

Our system uses vector embeddings to compute similarity between cultural entities, taking into consideration temporal alignment, relational proximity, and typological affinities. These computations involve a combination of two measures between the embedded graph structures: the cosine similarity, which works by calculating the cosine angle between two vector representations (though it is generally used for high dimensions dense vector comparisons), and the contrastive distance learning, which improves the clustering of semantically similar subgraphs and disperses the unrelated ones [117,118].
After passing three layers of graph convolution, the node embeddings encode rich enough structural and semantic information to perform tasks such as similarity computation: once the user query is transformed into a graph structure with the LLM, it is passed through the same CompGCN encoder to generate a comparable embedding in the same latent space as that of the knowledge graph. Cosine similarity is then computed between this query embedding and candidate entity or subgraph embeddings from the knowledge graph to find semantically relevant matches even in the absence of exact alignment (the candidates are scored in an interval of 0 to 1; 1 signifying high similarity). Figure 6 visualises this process into a simplified diagram.

6.5. Design and Implementation of the Prompt-Based Interface

Heritage Digital Twins are powerful archives, and alongside the extensive effort to improve the methods by which they are queried logically, these improvements must also be made accessible to users. The chosen approach aims to allow natural language as a querying medium, and as such, a prompt-based user interface was developed instead of the traditional, rigid hierarchy of filters or drop-down UIs. The concept is that of a chat environment for users to interact with the knowledge graph in a conversational manner. A key design consideration has been quality feedback and transparency. The model response must be synthesized and clear, allowing the user to understand how their prompt was parsed and interpreted. This is most valuable when the user’s query is ambiguous or lacking specificity. For example, a user query saying “Find me Greek figures from the classical period” would yield the response “Finding Greek historical personalities from the classical period. Would you like me to expand my search to mythological entities or sculptures and statues?”.
This points to another important element; iterative refinement. Following the model’s reasoning during the query would let the user, therefore, correct or refine the system’s interpretation of their intent. This requires the maintenance of a persistent conversation context that the system can always refer back to during the validity of dynamic session state, stored in an in-memory session store with a unique identifier passed with each message, after each of which the prompt context and history are updated and re-evaluated during the LLM convocations. The implementation of this feature further guarantees accessibility across expertise levels, as even unstructured queries can be evolved into their most effective form through a co-constructed exploration [119]. A well-rounded response is also often appreciated by non-experts if it provides additional information that might entice further interests in the topic at hand. For this, the LLM is also tasked to include adjacent context, or broader data summaries, informed by the knowledge graph into its response. This can be synthesised from the subgraphs scoring slightly lower in similarity against the query graph and should be clearly specified as such [119].

6.6. Presenting Complex Query Results in User-Comprehensible Formats

After a successful query, a sorted list of high-scoring entity or subgraph candidates for the given user prompt is obtained, and the task is then presented to the user in a human-readable layout. First, the LLM interprets the result data, searching for ways to highlight important information, categorise the entries, or provide contextual summaries where better suited; then, it returns it to the user in narrative format, as seen in Figure 7.
When dealing with large knowledge graphs such as the HDT’s, it is crucial to avoid clutter, filter through data, and present only the salient information initially. Progressive disclosure balances usability with transparency for the sake of both casual exploration and scholarly validation. Our current implementation of the UI presents the system’s response in a narrative format describing in plain text its findings, structured, listed, and categorised as needed by the LLM for better digestion of the information. Text-based responses are suitable for any type of queries, from precise search to exploratory inquisitions. They can be concise and direct but can also weave together complex temporal and geographic information, benefiting from explanation and nuance to paint a holistic picture for the user [120]. Not to suggest that text-based narration is the only format supported in chat-UIs; on the contrary, an important advantage is that they are extensible. In the context of HDTs, more structured or even visual elements would render the response more impactful, and thus the interface can be upgraded to include 3D model displays, images and illustrations, geographic and temporal maps, tables, diagrams, and raw graph structures depending on the user’s needs. Ultimately, presenting complex query results is not just a matter of UI feature design but a mediation between semantic accuracy and human cognition.

7. Discussion: Taming the Wild Beast

The hybrid querying approach described above clearly outperforms strict SPARQL translation in several important dimensions. First, it offers greater flexibility when dealing with user queries that are incomplete, underspecified, or expressed in natural language, such as “old buildings with mythological carvings”, where formal ontology mapping would struggle or fail entirely [121]. The system provides meaningful results even in the absence of precise predicates or labels by interpreting semantic intent and mapping it onto the knowledge graph structure instead of depending on exact term matching. Secondly, explainability is supported by the architecture. The query graph can be visualised, annotated, and even interactively refined by creating an intermediate graph representation for every query. Because it makes clear which entities and relations are being taken into consideration [102], this transparency aids users in understanding how their prompts are interpreted. Third, the method allows for gradual improvement. The system can mimic a human-like dialogue by asking clarifying follow-up questions, such as whether “Hellenistic” refers to a time period or an artistic style, in order to gradually refine the search instead of returning a null result because of strict schema constraints [77]. Moreover, the system empowers prompt engineering and context enrichment. A prompt like “Find ancient settlements near rivers active during early agriculture” can be interpreted by the LLM and enriched with relevant temporal ontologies and spatial reasoning constructs then structured into a graph-aware query. This graph is then embedded by CompGCN, which jointly encodes nodes and relations to support matching based on structure and context [93].

7.1. The Indispensable Role of Human Expertise in AI-Assisted Processes

Despite the growing capabilities of AI models to process, interpret, and retrieve information from both structured and unstructured data, human expertise remains a key part of any system that engages with rich cultural knowledge [101,122].
In AI-assisted querying over cultural heritage knowledge graphs, human input is not just an addition; it is essential for the reliability, interpretability, and contextual relevance of the whole system [101]. AI models, including prompt-based large language models (LLMs) and graph neural networks like CompGCN, are fundamentally statistical and generalising systems. Their ability to infer structure, relevance, and semantic similarity is limited by the quality and coverage of their training data and the ontological framing in the knowledge graph. Cultural heritage data is inherently varied, deeply contextual, and often only partially curated. This field includes ambiguities, historical gaps, and interpretive uncertainties that cannot fully be resolved by algorithms alone [101]. Human expertise is vital for curating representative datasets, designing prompt templates, aligning graph structures with domain ontologies like CIDOC CRM, and interpreting model outputs with historical context in mind [101,122]. Moreover, domain experts are crucial for assessing and improving the outputs of AI models. For example, interpreting a user query like “religious sites with serpent imagery” may involve understanding cultural motifs, symbolic meanings, and changes in iconography over time that are not clearly included in the ontology. In these situations, an expert is needed to confirm the system’s graph-based matches, identify semantically inappropriate connections, and help refine the query or results [101]. This ongoing feedback not only enhances the user experience but also supports the long-term improvement and strength of the AI model, especially when the feedback is included in training processes [122].
In the case of unavailable human expert validation, human input is still not disregarded. Cyclical processes in which the predictions or outputs of an AI model are evaluated by both humans and automated systems are called feedback loops [101,122] and are developed in AI systems to involve the user though feedback actions regarding AI results and suggestions. This feedback is turned into structured input that can enhance the underlying knowledge graph or fine-tune the model while also being used to evaluate performance metrics. In accepting, rejecting, or editing said recommendations, there is valuable human insight that the system translates into learning signals, and this information is captured as labeled training data.
The backend of the semantic pipeline can include modules for continuous or active learning, which allows the AI model to be retrained on new datasets that consider user input. Active learning techniques can help prioritise the most impactful or unclear cases for human review, making expert time more effective [101,122].
Error analysis is an important part of feedback loops. Developers can identify model blind spots or weaknesses in the training data by investigating systematic errors, such as frequent incorrect classifications or concepts that are not linked. In addition to model retraining, these insights can guide improvements to user instructions, ontology design, and data curation practices [101].
Additionally, explainability, which is vital for building trust in AI systems, relies on human interpretation. While neural architectures can generate relevance scores or attention weights, only experts can turn those signals into valid domain interpretations, such as explaining why a particular Hellenistic inscription appeared in response to a vaguely formulated mythological query [123]. Their knowledge helps bridge the gap between the model’s behaviour and human cultural understanding. Ultimately, the relationship between human expertise and machine intelligence creates a hybrid form of intelligence that is particularly effective in cultural heritage contexts. In this scenario, machines can enhance human reach and recall across extensive data collections, while humans provide interpretation, critique, and contextual relevance. This partnership ensures that AI tools do not simply replicate patterns but also contribute meaningfully to the advancement of knowledge, honoring both academic rigour and cultural sensitivity [124,125].

7.2. Ethical Implications and the Pursuit of Transparent AI in Cultural Heritage

As artificial intelligence systems play a larger role in accessing and interpreting cultural heritage data, important ethical issues need to be addressed, ensuring that these technologies uphold values such as transparency, accountability, and cultural sensitivity. Unlike traditional datasets, cultural heritage knowledge bases include narratives, identities, and power dynamics that reflect human histories and ways of knowing. Therefore, using AI in this area poses risks that go beyond technical bias, leading to issues such as historiographic distortion, cultural erasure, and unfair knowledge representation [126]. One major ethical challenge comes from the lack of clarity in current AI models, especially large language models and graph-based neural systems like CompGCN. While these systems can generate or retrieve rich context responses, they operate like “black boxes.” This makes it hard for users and curators to trace where a specific output comes from or to understand how a query match is made [127,128]. In cultural heritage applications, this lack of transparency is particularly concerning. Users may unknowingly receive results influenced by hidden biases in the data, such as Eurocentric narratives, underrepresentation of indigenous knowledge systems, or overly simplistic event interpretations outlined in systems like CIDOC CRM [126,129]. To reduce these risks, AI interfaces in this field should focus on transparency. This means developing interpretable structures, like graph-based representations used in query transformations, that users can visualise and annotate. Keeping a clear record of how data transforms from natural language input to query graph, and then to retrieved subgraphs, allows scholars and curators to review the model’s behavior and ensure accuracy [127]. Additionally, creating feedback systems that let users question or adjust AI outputs is crucial for maintaining human oversight and encouraging discussion [127]. Another key ethical issue is supporting diverse perspectives. Cultural heritage is not just a fixed set of facts; it is a dynamic field full of interpretation, often featuring contested meanings and various community viewpoints. AI models should be taught and assessed with awareness of these complexities. This involves using inclusive training materials, enabling multilingual and multicultural querying options, and making sure minority narratives are not oversimplified or ignored by dominant patterns [130,131]. Transparency in how models are developed, including clear documentation of data sources, design choices, and evaluation methods, helps prevent accidental cultural oversimplification [126,132]. Finally, using AI in this area must align with legal and ethical standards concerning data ownership and management. Cultural artefacts, especially those obtained through colonial means or held in sacred regard, require careful governance when they are digitized and processed by algorithms. Transparent and explainable AI should not only reveal its inner workings but also honor the rights and values of the communities whose heritage it engages with [126,133].

8. Conclusions and Future Work

The experiments described in this paper offer promising results, particularly in their capacity to simplify human tasks and provide tools that accelerate the extraction, organisation, and use of cultural knowledge. What emerges is not simply a more efficient pathway from text to data but a recursive system in which knowledge begets cognition, and cognition in turn refines the very tools through which knowledge is acquired. This methodological loop, anchored in ontological rigour and empowered by machine learning, reconfigures both the means and the ends of heritage data processing, transforming interpretation into interaction.
These outcomes represent a fundamental step towards the realisation of a fully developed Cognitive Heritage Digital Twin. As noted previously, providing a full account of the creation of an entire Cognitive Heritage Digital Twin would require a much larger study given the immense scale and complexity involved. The objective of this paper, therefore, was to exemplify some of the most immediately impactful trajectories currently emerging, centred on the use of artificial intelligence and ontologies for the retrieval, organisation, and strategic deployment of knowledge embedded in cultural heritage documentation. The presented AI-assisted semantic pipeline has been experimentally validated as a robust and efficient framework for knowledge extraction, laying the groundwork for more advanced functionalities. Future developments will aim to refine and extend the system by enhancing the accuracy of entity disambiguation, incorporating multimodal inputs such as images and tabular data, and introducing feedback loops that allow curators and domain experts to validate and iteratively improve the extracted knowledge while also enhancing the models being used.
As cultural heritage datasets span a broad spectrum of visual media, which is often heterogeneous, physically degraded, or embedded within ambiguous symbolic systems, it is important to also take those into account. Some promising zero-shot classification testing has already been conducted on images using open-source Vision Language Models like OPENClip [134] to automate the extraction of analysis types and semantic meaning to fuel Digital Twins and Knowledge Bases. This line of research will be further investigated to fine-tune the AI models and boost recognition of domain-specific figure types. Future plans will also aim to streamline the integration between natural language queries and graph-based responses, enabling more fluid and transparent interactions with the underlying knowledge base. These enhancements will not only increase the robustness of the extraction process but will further consolidate the cognitive capacities of Heritage Digital Twin, aligning their growth with the evolving needs of cultural heritage research. This entails not only refining the underlying semantic infrastructures but also designing more intuitive, transparent, and context-aware interfaces that allow scholars and practitioners to explore, interrogate, and contribute to the knowledge base with greater ease. By empowering both the machine and the human interlocutor, a more dialogic relationship with cultural data is envisioned, in which digital twins are no longer static repositories but dynamic environments for discovery, reasoning, and collaborative interpretation.
Particular attention will also be devoted to the applications of the AI–ontology tandem in the development of decisional components of Heritage Digital Twins, enabling them to react meaningfully to sensor-derived inputs and to activate appropriate, context-aware responses. Beyond decision-making and reactive behaviour, further functionalities will be explored, such as predictive modelling, simulation of possible conservation scenarios, and projection of future states of cultural assets under varying environmental or human-induced conditions. These developments will draw on the inferential and connective power of semantic graphs, which provide the epistemic infrastructure necessary to support such complex reasoning processes. This approach echoes the insight of Heraclitus, who warned that “much learning does not teach understanding” (Fragment B 40.). It is a powerful reminder, now as then, that true knowledge arises not from the mere accumulation of data but from its structuring, contextualisation, and meaningful integration. Our work constantly strives to answer this call: to move beyond erudition towards insight; beyond cataloguing towards comprehension.
Naturally, all future enhancements to the pipeline and to the cognitive architecture of the Heritage Digital Twin will be guided by rigorous ethical considerations and by a persistent commitment to human oversight. As artificial intelligence systems become more sophisticated in navigating and reconstructing cultural knowledge, their design must remain transparent, verifiable, and accountable. As narrated at the beginning of this paper, the gods fashioned a simulacrum, identical to the real Helen in form, voice, and grace, to deceive mortals and at the same time to preserve and keep her safe from the violence of men. Digital twins too are conceived to preserve and protect, but certainly not to deceive, since by definition their raison d’être lies not in illusion but in historical and epistemic fidelity. Yet in pursuing verisimilitude and cognitive depth, the risk is of being misled by the very intelligences we have created. Artificial systems, however powerful, often yield responses that are incomplete, imprecise, or altogether fabricated, i.e., hallucinations mistaken for knowledge. The risk, then, is that we ourselves are deceived by the AI meant to assist us, subtly shifting from curators of memory to unwitting custodians of fictions. The future of Heritage Digital Twins thus resides at the intersection of culture, computation, and semantics [135], transforming them into dynamic epistemological machines capable of simulating worlds, diagnosing faults, predicting futures, and structuring meaning. From this vision emerges a new notion of the Cognitive Digital Twin: a system capable of transcending passive simulation to engage in active reasoning, hypothesis generation, and adaptive control [136].

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/app151810061/s1.

Author Contributions

Investigation, A.F., A.H. and M.S. All authors have equally contributed to the present paper. All authors have read and agreed to the published version of the manuscript.

Funding

This paper received no external funding.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

The data presented in this study are available on request from the corresponding author. The data are not publicly available due to privacy restrictions.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Graves, R. The Greek Myths; Penguin Books: London, UK, 1955; Volume 2, pp. 274–275. [Google Scholar]
  2. Jones, D.; Snider, C.; Nassehi, A.; Yon, J.; Hicks, B. Characterising the Digital Twin: A systematic literature review. CIRP J. Manuf. Sci. Technol. 2020, 29, 36–52. [Google Scholar] [CrossRef]
  3. Semeraro, C.; Lezoche, M.; Panetto, H.; Dassisti, M. Digital twin paradigm: A systematic literature review. Comput. Ind. 2021, 130, 103469. [Google Scholar] [CrossRef]
  4. Schrotter, G.; Hürzeler, C. The Digital Twin of the City of Zurich for Urban Planning. PFG—J. Photogramm. Remote Sens. Geoinf. Sci. 2020, 88, 99–112. [Google Scholar] [CrossRef]
  5. Caprari, G.; Castelli, G.; Montuori, M.; Camardelli, M.; Malvezzi, R. Digital Twin for Urban Planning in the Green Deal Era: A State of the Art and Future Perspectives. Sustainability 2022, 14, 6263. [Google Scholar] [CrossRef]
  6. Grieves, M.; Vickers, J. Digital Twin: Mitigating Unpredictable, Undesirable Emergent Behavior in Complex Systems. In Transdisciplinary Perspectives on Complex Systems: New Findings and Approaches; Kahlen, F.J., Flumerfelt, S., Alves, A., Eds.; Springer International Publishing: Cham, Switzerland, 2017; pp. 85–113. [Google Scholar] [CrossRef]
  7. Li, L.; Aslam, S.; Wileman, A.; Perinpanayagam, S. Digital Twin in Aerospace Industry: A Gentle Introduction. IEEE Access 2022, 10, 9543–9562. [Google Scholar] [CrossRef]
  8. Alfaro-Viquez, D.; Zamora-Hernandez, M.; Fernandez-Vega, M.; Garcia-Rodriguez, J.; Azorin-Lopez, J. A Comprehensive Review of AI-Based Digital Twin Applications in Manufacturing: Integration Across Operator, Product, and Process Dimensions. Electronics 2025, 14, 646. [Google Scholar] [CrossRef]
  9. Bunea, A.A.; Dobre, C.M. Developing an Immersive Virtual Reality Interface for Urban Digital Twins: Integrating CAVE and HMD Systems for Enhanced Urban Analysis. In Proceedings of the 2024 23rd RoEduNet Conference: Networking in Education and Research (RoEduNet), Bucharest, Romania, 19–20 September 2024; pp. 1–5. [Google Scholar] [CrossRef]
  10. European Commission. Destination Earth. Available online: https://destination-earth.eu (accessed on 15 July 2025).
  11. Nikolinakos, N.T. EU Policy and Legal Framework for Artificial Intelligence, Robotics and Related Technologies—The AI Act, 1st ed.; Number v.53 in Law, Governance and Technology Series; Springer International Publishing AG: Cham, Switzerland, 2023. [Google Scholar]
  12. Niccolucci, F.; Felicetti, A.; Hermon, S. Populating the Data Space for Cultural Heritage with Heritage Digital Twins. Data 2022, 7, 105. [Google Scholar] [CrossRef]
  13. Niccolucci, F.; Markhoff, B.; Theodoridou, M.; Felicetti, A.; Hermon, S. The Heritage Digital Twin: A bicycle made for two. The integration of digital methodologies into cultural heritage research. Open Res. Eur. 2023, 3, 64. [Google Scholar] [CrossRef]
  14. Niccolucci, F.; Felicetti, A. Digital Twin Sensors in Cultural Heritage Ontology Applications. Sensors 2024, 24, 3978. [Google Scholar] [CrossRef]
  15. Hedrick, E.; Hatch, D.; Oliver, E. Digital Twins in the Ancient and Historical World. In Proceedings of the 2025 Intermountain Engineering, Technology and Computing (IETC), Orem, UT, USA, 9–10 May 2025; pp. 1–6. [Google Scholar] [CrossRef]
  16. Luther, W.; Baloian, N.; Biella, D.; Sacher, D. Digital Twins and Enabling Technologies in Museums and Cultural Heritage: An Overview. Sensors 2023, 23, 1583. [Google Scholar] [CrossRef]
  17. Menaguale, O. Digital Twin and Cultural Heritage—The Future of Society Built on History and Art. In The Digital Twin; Crespi, N., Drobot, A.T., Minerva, R., Eds.; Springer International Publishing: Cham, Switzerland, 2023; pp. 1081–1111. [Google Scholar] [CrossRef]
  18. Amelio, A.; Zarri, G.P. Cultural heritage digital twin: Modeling and representing the visual narrative in Leonardo Da Vinci’s Mona Lisa. Neural Comput. Appl. 2024, 36, 11859–11876. [Google Scholar] [CrossRef]
  19. Gros, A.; Guillem, A.; De Luca, L.; Baillieul, É.; Duvocelle, B.; Malavergne, O.; Leroux, L.; Zimmer, T. Faceting the post-disaster built heritage reconstruction process within the digital twin framework for Notre-Dame de Paris. Sci. Rep. 2023, 13, 5981. [Google Scholar] [CrossRef]
  20. Ni, Z.; Eriksson, P.; Liu, Y.; Karlsson, M.; Gong, S. Improving energy efficiency while preserving historic buildings with digital twins and artificial intelligence. IOP Conf. Ser. Earth Environ. Sci. 2021, 863, 012041. [Google Scholar] [CrossRef]
  21. Ni, Z.; Hupkes, J.; Eriksson, P.; Leijonhufvud, G.; Karlsson, M.; Gong, S. Parametric Digital Twins for Preserving Historic Buildings: A Case Study at Löfstad Castle in Östergötland, Sweden. IEEE Access 2025, 13, 3371–3389. [Google Scholar] [CrossRef]
  22. Zhang, Z.; Dang, A.; Huang, J.; Chen, Y. Advancing Conservation Methods of the Great Wall Cultural Heritage Through Digital Twin. IEEE Internet Comput. 2025, 29, 48–55. [Google Scholar] [CrossRef]
  23. Brandt, A.K. Beethoven’s Ninth and AI’s Tenth: A comparison of human and computational creativity. J. Creat. 2023, 33, 100068. [Google Scholar] [CrossRef]
  24. Rizk, R.; Rizk, D.; Rizk, F.; Kumar, A. A Hybrid Capsule Network-based Deep Learning Framework for Deciphering Ancient Scripts with Scarce Annotations: A Case Study on Phoenician Epigraphy. In Proceedings of the 2021 IEEE International Midwest Symposium on Circuits and Systems (MWSCAS), Lansing, MI, USA, 9–11 August 2021; pp. 617–620. [Google Scholar] [CrossRef]
  25. Vats, S.; Mehta, S. Revolutionizing Archaeological Discoveries: The Role of Artificial Intelligence and Machine Learning in Site Analysis. In Proceedings of the 2024 15th International Conference on Computing Communication and Networking Technologies (ICCCNT), Kamand, India, 24–28 June 2024; pp. 1–5. [Google Scholar] [CrossRef]
  26. Cobb, P.J. Large Language Models and Generative AI, Oh My!: Archaeology in the Time of ChatGPT, Midjourney, and Beyond. Adv. Archaeol. Pract. 2023, 11, 363–369. [Google Scholar] [CrossRef]
  27. Leong, W.Y. Revitalizing Historical Paintings through Artificial Intelligence. In Proceedings of the 2025 Joint International Conference on Digital Arts, Media and Technology with ECTI Northern Section Conference on Electrical, Electronics, Computer and Telecommunications Engineering (ECTI DAMT & NCON), Nan, Thailand, 29 January–1 February 2025; pp. 141–145. [Google Scholar] [CrossRef]
  28. Pisu, F.A.; Porcu, S.; Carboni, R.; Mameli, V.; Cannas, C.; Naitza, S.; Melis, R.T.; Mussi, M.; Chiriu, D. Innovative method for provenance studies in cultural heritage: A new algorithm based on observables from high-resolution Raman spectra of red ochre. Spectrochim. Acta Part A Mol. Biomol. Spectrosc. 2025, 329, 125581. [Google Scholar] [CrossRef]
  29. Mishra, M.; Barman, T.; Ramana, G. Artificial intelligence-based visual inspection system for structural health monitoring of cultural heritage. J. Civ. Struct. Health Monit. 2024, 14, 103–120. [Google Scholar] [CrossRef]
  30. Ararat, M.B.; Yenice, T.K. The use of artificial intelligence in damage assessment of historical buildings. In Frontiers of Architectural Research; Elsevier: Amsterdam, The Netherlands, 2025. [Google Scholar]
  31. Gaber, J.A.; Youssef, S.M.; Fathalla, K.M. The role of artificial intelligence and machine learning in preserving cultural heritage and art works via virtual restoration. ISPRS Ann. Photogramm. Remote Sens. Spat. Inf. Sci. 2023, 10, 185–190. [Google Scholar] [CrossRef]
  32. Fuller, A.; Fan, Z.; Day, C.; Barlow, C. Digital Twin: Enabling Technologies, Challenges and Open Research. IEEE Access 2020, 8, 108952–108971. [Google Scholar] [CrossRef]
  33. Mayer, A.; Greif, L.; Häußermann, T.M.; Otto, S.; Kastner, K.; El Bobbou, S.; Chardonnet, J.R.; Reichwald, J.; Fleischer, J.; Ovtcharova, J. Digital Twins, Extended Reality, and Artificial Intelligence in Manufacturing Reconfiguration: A Systematic Literature Review. Sustainability 2025, 17, 2318. [Google Scholar] [CrossRef]
  34. Charalabidis, Y.; Kontos, G.; Zitianellis, D. Local Digital Twins for Cities and Regions: The Way Forward. In Decide Better: Open and Interoperable Local Digital Twins; Raes, L., Ruston McAleer, S., Croket, I., Kogut, P., Brynskov, M., Lefever, S., Eds.; Springer Nature Switzerland: Cham, Switzerland, 2025; pp. 233–259. [Google Scholar] [CrossRef]
  35. Tampère, C.; Ortmann, P.; Jedlička, K.; Lohman, W.; Janssen, S. Future Ready Local Digital Twins and the Use of Predictive Simulations: The Case of Traffic and Traffic Impact Modelling. In Decide Better: Open and Interoperable Local Digital Twins; Raes, L., Ruston McAleer, S., Croket, I., Kogut, P., Brynskov, M., Lefever, S., Eds.; Springer Nature Switzerland: Cham, Switzerland, 2025; pp. 203–230. [Google Scholar] [CrossRef]
  36. Arsalan, H.; Heesom, D.; Moore, N. From Heritage Building Information Modelling Towards an ‘Echo-Based’ Heritage Digital Twin. Heritage 2025, 8, 33. [Google Scholar] [CrossRef]
  37. Fan, C.; Zhang, C.; Yahja, A.; Mostafavi, A. Disaster City Digital Twin: A vision for integrating artificial and human intelligence for disaster management. Int. J. Inf. Manag. 2021, 56, 102049. [Google Scholar] [CrossRef]
  38. Lucchi, E. Digital twins, artificial intelligence and immersive technologies for heritage preservation and cultural tourism in smart cities. In Digital Twin, Blockchain, and Sensor Networks in the Healthy and Mobile City; Elsevier: Amsterdam, The Netherlands, 2025; pp. 507–520. [Google Scholar]
  39. Pavlidis, G. Agentic AI for Cultural Heritage: Embedding Risk Memory in Semantic Digital Twins. Computers 2025, 14, 266. [Google Scholar] [CrossRef]
  40. Singh, M.; Fuenmayor, E.; Hinchy, E.; Qiao, Y.; Murray, N.; Devine, D. Digital Twin: Origin to Future. Appl. Syst. Innov. 2021, 4, 36. [Google Scholar] [CrossRef]
  41. Leipnitz, M.T.; Petry, R.H.; Rodrigues, F.H.; De Souza Silva, H.R.; Correia, J.B.; Becker, K.; Wickboldt, J.A.; Carbonera, J.L.; Netto, J.C.; Abel, M. Architecting Digital Twins: From Concept to Reality. Procedia Comput. Sci. 2025, 256, 530–537. [Google Scholar] [CrossRef]
  42. Rathore, M.M.; Shah, S.A.; Shukla, D.; Bentafat, E.; Bakiras, S. The Role of AI, Machine Learning, and Big Data in Digital Twinning: A Systematic Literature Review, Challenges, and Opportunities. IEEE Access 2021, 9, 32030–32052. [Google Scholar] [CrossRef]
  43. Boje, C.; Guerriero, A.; Kubicki, S.; Rezgui, Y. Towards a semantic Construction Digital Twin: Directions for future research. Autom. Constr. 2020, 114, 103179. [Google Scholar] [CrossRef]
  44. Karabulut, E.; Pileggi, S.F.; Groth, P.; Degeler, V. Ontologies in digital twins: A systematic literature review. Future Gener. Comput. Syst. 2024, 153, 442–456. [Google Scholar] [CrossRef]
  45. Xhani, D.; Moreira, J.L.R.; van Sinderen, M.; Pires, L.F. XMLPO: An Ontology for Explainable Machine Learning Pipeline. In Frontiers in Artificial Intelligence and Applications; IOS Press: Amsterdam, The Netherlands, 2024; pp. 193–206. ISSN 0922-6389. [Google Scholar] [CrossRef]
  46. Albers, L.; Große, P.; Wagner, S. Semantic Data-Modeling and Long-Term Interpretability of Cultural Heritage Data—Three Case Studies. In Digital Cultural Heritage; Springer International Publishing: Cham, Switzerland, 2020; pp. 239–253. [Google Scholar] [CrossRef]
  47. Felicetti, A.; Meghini, C.; Richards, J.; Theodoridou, M. The AO-Cat Ontology. Technical Report, Zenodo, 2025. Version Number: 1.2. [CrossRef]
  48. CIDOC CRM. Available online: https://cidoc-crm.org (accessed on 10 July 2025).
  49. ISO 21127:2023; Information and documentation—A Reference Ontology for the Interchange of Cultural Heritage Information. ISO: Geneva, Switzerland, 2023.
  50. CRMsci—Scientific Observation Model. Available online: https://cidoc-crm.org/crmsci (accessed on 11 July 2025).
  51. CRMdig—Model for Provenance Metadata. Available online: https://cidoc-crm.org/crmdig (accessed on 14 July 2025).
  52. Castelli, L.; Felicetti, A.; Proietti, F. Heritage Science and Cultural Heritage: standards and tools for establishing cross-domain data interoperability. Int. J. Digit. Libr. 2021, 22, 279–287. [Google Scholar] [CrossRef]
  53. Felicetti, A.; Niccolucci, F. Artificial Intelligence and Ontologies for the Management of Heritage Digital Twins Data. Data 2024, 10, 1. [Google Scholar] [CrossRef]
  54. Alammar, J.; Grootendorst, M. Hands-On Large Language Models: Language Understanding and Generation, 1st ed.; O’Reilly Media, Inc.: Sebastopol, CA, USA, 2024. [Google Scholar]
  55. Chua, L.; Ghazi, B.; Huang, Y.; Kamath, P.; Kumar, R.; Manurangsi, P.; Sinha, A.; Xie, C.; Zhang, C. Crosslingual Capabilities and Knowledge Barriers in Multilingual Large Language Models. arXiv 2025, arXiv:2406.16135. [Google Scholar]
  56. Yang, H.; Zhang, Y.; Xu, J.; Lu, H.; Heng, P.A.; Lam, W. Unveiling the Generalization Power of Fine-Tuned Large Language Models. arXiv 2024, arXiv:2403.09162. [Google Scholar] [CrossRef]
  57. Schick, T.; Schütze, H. Exploiting Cloze Questions for Few Shot Text Classification and Natural Language Inference. arXiv 2021, arXiv:2001.07676. [Google Scholar] [CrossRef]
  58. Tudor, C.; Megyesi, B.; Östling, R. Prompting the Past: Exploring Zero-Shot Learning for Named Entity Recognition in Historical Texts Using Prompt-Answering LLMs. In Proceedings of the 9th Joint SIGHUM Workshop on Computational Linguistics for Cultural Heritage, Social Sciences, Humanities and Literature (LaTeCH-CLfL 2025), Albuquerque, NM, USA, 3–4 May 2025; Kazantseva, A., Szpakowicz, S., Degaetano-Ortlieb, S., Bizzoni, Y., Pagel, J., Eds.; pp. 216–226. [Google Scholar] [CrossRef]
  59. Brown, N.B. Enhancing Trust in LLMs: Algorithms for Comparing and Interpreting LLMs. arXiv 2024, arXiv:2406.01943. [Google Scholar] [CrossRef]
  60. Cullhed, E. Instruct-Tuning Pretrained Causal Language Models for Ancient Greek Papyrology and Epigraphy. arXiv 2024, arXiv:2409.13870. [Google Scholar]
  61. Ignatowicz, J.; Kutt, K.; Nalepa, G.J. Position Paper: Metadata Enrichment Model: Integrating Neural Networks and Semantic Knowledge Graphs for Cultural Heritage Applications. arXiv 2025, arXiv:2505.23543. [Google Scholar] [CrossRef]
  62. Meoded, E. Handwritten Text Recognition of Historical Manuscripts Using Transformer-Based Models. arXiv 2025, arXiv:2508.11499. [Google Scholar] [CrossRef]
  63. Zhang, X.; Bos, J. Multi-Modal Semantic Parsing for the Interpretation of Tombstone Inscriptions. arXiv 2025, arXiv:2507.04377. [Google Scholar] [CrossRef]
  64. Schweter, S.; März, L.; Schmid, K.; Çano, E. hmBERT: Historical Multilingual Language Models for Named Entity Recognition. arXiv 2022, arXiv:2205.15575. [Google Scholar] [CrossRef]
  65. ArchaeoBERT. Available online: https://huggingface.co/alexbrandsen/ArchaeoBERT (accessed on 20 August 2025).
  66. Ehrmann, M.; Hamdi, A.; Pontes, E.L.; Romanello, M.; Doucet, A. Named Entity Recognition and Classification in Historical Documents: A Survey. ACM Comput. Surv. 2023, 56, 1–47. [Google Scholar] [CrossRef]
  67. Sprugnoli, R.; Tonelli, S. Novel Event Detection and Classification for Historical Texts. Comput. Linguist. 2019, 45, 229–265. [Google Scholar] [CrossRef]
  68. Zhu, Y.; Wang, X.; Chen, J.; Qiao, S.; Ou, Y.; Yao, Y.; Deng, S.; Chen, H.; Zhang, N. LLMs for Knowledge Graph Construction and Reasoning: Recent Capabilities and Future Opportunities. arXiv 2024, arXiv:2305.13168. [Google Scholar] [CrossRef]
  69. Yang, L.; Chen, H.; Li, Z.; Ding, X.; Wu, X. Give Us the Facts: Enhancing Large Language Models with Knowledge Graphs for Fact-aware Language Modeling. arXiv 2024, arXiv:2306.11489. [Google Scholar] [CrossRef]
  70. Tang, N.; Fan, J.; Li, F.; Tu, J.; Du, X.; Li, G.; Madden, S.; Ouzzani, M. RPT: Relational Pre-trained Transformer Is Almost All You Need towards Democratizing Data Preparation. arXiv 2021, arXiv:2012.02469. [Google Scholar] [CrossRef]
  71. Amer-Yahia, S.; Bonifati, A.; Chen, L.; Li, G.; Shim, K.; Xu, J.; Yang, X. From Large Language Models to Databases and Back: A discussion on research and education. arXiv 2023, arXiv:2306.01388. [Google Scholar] [CrossRef]
  72. Choi, Y.; Asif, M.A.; Han, Z.; Willes, J.; Krishnan, R.G. Teaching LLMs How to Learn with Contextual Fine-Tuning. arXiv 2025, arXiv:2503.09032. [Google Scholar] [CrossRef]
  73. Pyrih, V.; Rebmann, A.; van der Aa, H. LLMs that Understand Processes: Instruction-tuning for Semantics-Aware Process Mining. arXiv 2025, arXiv:2508.16270. [Google Scholar]
  74. Jiménez-Ruiz, E.; Grau, B. LogMap: Logic-Based and Scalable Ontology Matching; Springer: Berlin/Heidelberg, Germany, 2011; pp. 273–288. [Google Scholar] [CrossRef]
  75. He, Y.; Chen, J.; Antonyrajah, D.; Horrocks, I. BERTMap: A BERT-based Ontology Alignment System. arXiv 2022, arXiv:2112.02682. [Google Scholar] [CrossRef]
  76. Fensel, D.; Şimşek, U.; Angele, K.; Huaman, E.; Kärle, E.; Panasiuk, O.; Toma, I.; Umbrich, J.; Wahler, A. How to Build a Knowledge Graph. In Knowledge Graphs: Methodology, Tools and Selected Use Cases; Springer International Publishing: Cham, Switzerland, 2020; pp. 11–68. [Google Scholar] [CrossRef]
  77. Maree, M. Quantifying Relational Exploration in Cultural Heritage Knowledge Graphs with LLMs: A Neuro-Symbolic Approach for Enhanced Knowledge Discovery. Data 2025, 10, 52. [Google Scholar] [CrossRef]
  78. Bruns, O. The Persistence of Temporality: Representation of Time in Cultural Heritage Knowledge Graphs. In Proceedings of the Doctoral Consortium at ISWC 2023, co-located with 22nd International Semantic Web Conference (ISWC 2023), Athens, Greece, 7 November 2023; ISWC 2023. Volume 3678. [Google Scholar]
  79. Faralli, S.; Lenzi, A.; Velardi, P. A Large Interlinked Knowledge Graph of the Italian Cultural Heritage. In Proceedings of the LREC, Marseille, France, 20–25 June 2022; pp. 6280–6289. [Google Scholar]
  80. Carriero, V.A.; Gangemi, A.; Mancinelli, M.L.; Marinucci, L.; Nuzzolese, A.G.; Presutti, V.; Veninata, C. ArCo: The Italian Cultural Heritage Knowledge Graph. arXiv 2019, arXiv:1905.02840. [Google Scholar] [CrossRef]
  81. Barzaghi, S.; Moretti, A.; Heibi, I.; Peroni, S. CHAD-KG: A Knowledge Graph for Representing Cultural Heritage Objects and Digitisation Paradata. arXiv 2025, arXiv:2505.13276. [Google Scholar] [CrossRef]
  82. Devlin, J.; Chang, M.W.; Lee, K.; Toutanova, K. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. arXiv 2018, arXiv:1810.04805. [Google Scholar] [CrossRef]
  83. Brown, T.B.; Mann, B.; Ryder, N.; Subbiah, M.; Kaplan, J.; Dhariwal, P.; Neelakantan, A.; Shyam, P.; Sastry, G.; Askell, A.; et al. Language Models are Few-Shot Learners. arXiv 2020, arXiv:2005.14165. [Google Scholar] [CrossRef]
  84. Touvron, H.; Lavril, T.; Izacard, G.; Martinet, X.; Lachaux, M.A.; Lacroix, T.; Rozière, B.; Goyal, N.; Hambro, E.; Azhar, F.; et al. LLaMA: Open and Efficient Foundation Language Models. arXiv 2023, arXiv:2302.13971. [Google Scholar] [CrossRef]
  85. Gekhman, Z.; Yona, G.; Aharoni, R.; Eyal, M.; Feder, A.; Reichart, R.; Herzig, J. Does Fine-Tuning LLMs on New Knowledge Encourage Hallucinations? arXiv 2024, arXiv:2405.05904. [Google Scholar] [CrossRef]
  86. Pan, S.; Luo, L.; Wang, Y.; Chen, C.; Wang, J.; Wu, X. Unifying Large Language Models and Knowledge Graphs: A Roadmap. IEEE Trans. Knowl. Data Eng. 2024, 36, 3580–3599. [Google Scholar] [CrossRef]
  87. Agrawal, A.; Suzgun, M.; Mackey, L.; Kalai, A.T. Do Language Models Know When They’re Hallucinating References? arXiv 2023, arXiv:2305.18248. [Google Scholar] [CrossRef]
  88. Sui, Y.; He, Y.; Ding, Z.; Hooi, B. Can Knowledge Graphs Make Large Language Models More Trustworthy? An Empirical Study Over Open-ended Question Answering. arXiv 2025, arXiv:2410.08085. [Google Scholar] [CrossRef]
  89. Sun, J.; Du, Z.; Chen, Y. Knowledge Graph Tuning: Real-time Large Language Model Personalization based on Human Feedback. arXiv 2024, arXiv:2405.19686. [Google Scholar] [CrossRef]
  90. Doerr, M. The CIDOC Conceptual Reference Module: An Ontological Approach to Semantic Interoperability of Metadata. AI Mag. 2003, 24, 75–92. [Google Scholar]
  91. Binding, C.; May, K.; Tudhope, D. Semantic interoperability in archaeological datasets: Data mapping and extraction via the CIDOC CRM. In Proceedings of the Research and Advanced Technology for Digital Libraries: 15th International Conference on Theory and Practice of Digital Libraries (TPDL 2011), Berlin, Germany, 23–26 September 2011; Springer: Berlin, Germany, 2011; pp. 280–291. [Google Scholar]
  92. Doerr, M.; Kritsotaki, A.; Rousakis, Y. CIDOC2VEC: Extracting Information from Atomized CIDOC-CRM Humanities Knowledge Graphs. Information 2021, 12, 503. [Google Scholar]
  93. Vashishth, S.; Sanyal, S.; Nitin, V.; Talukdar, P. Composition-based multi-relational graph convolutional networks. In Proceedings of the International Conference on Learning Representations (ICLR), Addis Ababa, Ethiopia, 26–30 April 2020. [Google Scholar]
  94. Schlichtkrull, M.; Kipf, T.N.; Bloem, P.; van den Berg, R.; Titov, I.; Welling, M. Modeling relational data with graph convolutional networks. In Proceedings of the European Semantic Web Conference (ESWC), Crete, Greece, 3–7 June 2018; Springer: Berlin, Germany, 2018; pp. 593–607. [Google Scholar]
  95. El Vaigh, C.B.; Garcia, N.; Renoust, B.; Chu, C.; Nakashima, Y.; Nagahara, H. GCNBoost: Artwork Classification by Label Propagation through a Knowledge Graph. arXiv 2021, arXiv:2105.11852. [Google Scholar] [CrossRef]
  96. Meroño-Peñuela, A.; Sack, H. Move Cultural Heritage Knowledge Graphs in Everyone’s Pocket. Semant. Web J. 2020, 14, 323–359. [Google Scholar]
  97. Yuan, H.; Li, Y.; Wang, B.; Liu, K.; Zhang, J. Knowledge Graph-Based Intelligent Question Answering System for Ancient Chinese Costume Heritage. NPJ Herit. Sci. 2025, 13, 198. [Google Scholar] [CrossRef]
  98. Huang, Y.; Yu, S.; Chu, J.; Fan, H.; Du, B. Using Knowledge Graphs and Deep Learning Algorithms to Enhance Digital Cultural Heritage Management. Herit. Sci. 2023, 11, 204. [Google Scholar] [CrossRef]
  99. Ji, S.; Pan, S.; Cambria, E.; Marttinen, P.; Yu, P.S. A Survey on Knowledge Graphs: Representation, Acquisition, and Applications. IEEE Trans. Neural Netw. Learn. Syst. 2021, 33, 494–514. [Google Scholar] [CrossRef]
  100. Majumder, E.A. Injecting Knowledge Graphs into Large Language Models. arXiv 2025, arXiv:2505.07554. [Google Scholar] [CrossRef]
  101. Zhang, B.; Meroño Peñuela, A.; Simperl, E. Towards Explainable Automatic Knowledge Graph Construction with Human-in-the-loop. Semant. Web J. 2023, 274–289. [Google Scholar] [CrossRef]
  102. Arazzi, M.; Ligari, D.; Nicolazzo, S.; Nocera, A. Augmented Knowledge Graph Querying leveraging LLMs. arXiv 2025, arXiv:2502.01298. [Google Scholar]
  103. Wang, J.; Cui, Z.; Wang, B.; Pan, S.; Gao, J.; Yin, B.; Gao, W. IME: Integrating Multi-curvature Shared and Specific Embedding for Temporal Knowledge Graph Completion. arXiv 2024, arXiv:2403.19881. [Google Scholar] [CrossRef]
  104. Gaşi, A.; Ensari, T.; Dağtekin, M. Ontology supported semantic based image retrieval. Semant. Web. 2024, 15, 2125–2137. [Google Scholar] [CrossRef]
  105. Castellano, G.; Sansaro, G.; Vessio, G. Integrating Contextual Knowledge to Visual Features for Fine Art Classification. arXiv 2021, arXiv:2105.15028. [Google Scholar] [CrossRef]
  106. ARTEMIS. Applying Reactive Twins to Enhance Monument Information Systems. Available online: https://www.artemis-twin.eu (accessed on 12 July 2025).
  107. PyMuPDF. Available online: https://pymupdf.readthedocs.io/en/latest/ (accessed on 20 August 2025).
  108. SpaCy. Available online: https://spacy.io/ (accessed on 20 August 2025).
  109. All-MiniLM-L6-v2. Available online: https://huggingface.co/sentence-transformers/all-MiniLM-L6-v2 (accessed on 20 August 2025).
  110. Mistral 7B Instruct. Available online: https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.3 (accessed on 20 August 2025).
  111. FAISS. Available online: https://faiss.ai/index.html (accessed on 20 August 2025).
  112. The AO-Cat Ontology. Available online: https://zenodo.org/records/7818375 (accessed on 14 July 2025).
  113. ARIADNE Research Infrastructure. Available online: https://www.ariadne-research-infrastructure.eu (accessed on 12 July 2025).
  114. Mazzinghi, A.; Ruberto, C.; Giuntini, L.; Mandò, P.A.; Taccetti, F.; Castelli, L. Mapping with Macro X-ray Fluorescence Scanning of Raffaello’s Portrait of Leo X. Heritage 2022, 5, 3993–4005. [Google Scholar] [CrossRef]
  115. Leboef, P.; Doerr, M. CIDOC Conceptual Reference Model: A Standard for Cultural Heritage Information. J. Comput. Cult. Herit. 2017, 10, 1–20. [Google Scholar]
  116. Bordes, A.; Usunier, N.; Garcia-Duran, A.; Weston, J.; Yakhnenko, O. Translating Embeddings for Modeling Multi-relational Data. Adv. Neural Inf. Process. Syst. 2013, 26. [Google Scholar]
  117. Xu, D.; Cheng, W.; Luo, D.; Chen, H.; Zhang, X. InfoGCL: Information-Aware Graph Contrastive Learning. arXiv 2021, arXiv:2110.15438. [Google Scholar]
  118. Liang, K.; Liu, Y.; Zhou, S.; Tu, W.; Wen, Y.; Yang, X.; Dong, X.; Liu, X. Knowledge Graph Contrastive Learning Based on Relation-Symmetrical Structure. arXiv 2022, arXiv:2211.10738. [Google Scholar] [CrossRef]
  119. Füßl, A.; Nissen, V. Design of Explanation User Interfaces for Interactive Machine Learning. Requir. Eng. 2025. [Google Scholar] [CrossRef]
  120. Barzaghi, S.; Colitti, S.; Moretti, A.; Renda, G. From Metadata to Storytelling: A Framework For 3D Cultural Heritage Visualization on RDF Data. arXiv 2025, arXiv:2505.14328. [Google Scholar] [CrossRef]
  121. Emonet, V.; Bolleman, J.; Duvaud, S.; de Farias, T.; Sima, A.C. LLM-based SPARQL Query Generation from Natural Language over Federated Knowledge Graphs. arXiv 2025, arXiv:2410.06062. [Google Scholar]
  122. Manzoor, E.; Tong, J.; Vijayaraghavan, S.; Li, R. Expanding Knowledge Graphs with Humans in the Loop. arXiv 2023, arXiv:2212.05189. [Google Scholar] [CrossRef]
  123. Celino, I. Who is this Explanation for? Human Intelligence and Knowledge Graphs for eXplainable AI. arXiv 2020, arXiv:2005.13275. [Google Scholar] [CrossRef]
  124. Kamar, E. Directions in Hybrid Intelligence: Complementing AI Systems with Human Intelligence. In Proceedings of the Twenty-Fifth International Joint Conference on Artificial Intelligence, New York, NY, USA, 9–15 July 2016. [Google Scholar]
  125. Te’eni, D.; Yahav, I.; Zagalsky, A.; Schwartz, D.G.; Silverman, G. Reciprocal Human—Machine Learning: A Theory and an Instantiation for the Case of Message Classification. Manag. Sci. 2023. [Google Scholar] [CrossRef]
  126. Foka, A.; Griffin, G.; Ortiz Pablo, D.; Rajkowska, P.; Badri, S. Tracing the bias loop: AI, cultural heritage and bias-mitigating in practice. Ai Soc. 2025. [Google Scholar] [CrossRef]
  127. Tóth, G.M.; Albrecht, R.; Pruski, C. Explainable AI, LLM, and digitized archival cultural heritage: A case study of the Grand Ducal Archive of the Medici. AI Soc. 2025. [Google Scholar] [CrossRef]
  128. Peters, U.; Carman, M. Cultural Bias in Explainable AI Research: A Systematic Analysis. J. Artif. Intell. Res. 2024. [Google Scholar] [CrossRef]
  129. Mitchell, M. AI Is Spreading Old Stereotypes to New Languages and Cultures. Wired 2025. [Google Scholar]
  130. Upmann, P. What Role Does Cultural Context Play in Defining Ethical Standards for AI? AI Glob. 2024. [Google Scholar]
  131. Prabhakaran, V.; Qadri, R.; Hutchinson, B. Cultural Incongruencies in Artificial Intelligence. arXiv 2022, arXiv:2211.13069. [Google Scholar] [CrossRef]
  132. Inuwa-Dutse, I. FATE in AI: Towards Algorithmic Inclusivity and Accessibility. arXiv 2023, arXiv:2301.01590. [Google Scholar] [CrossRef]
  133. Kukutai, T.; Taylor, J. Indigenous Data Sovereignty: Towards an Agenda; ANU Press: Canberra, Australia, 2020. [Google Scholar]
  134. Cherti, M.; Beaumont, R.; Wightman, R.; Wortsman, M.; Ilharco, G.; Gordon, C.; Schuhmann, C.; Schmidt, L.; Jitsev, J. Reproducible Scaling Laws for Contrastive Language-Image Learning. In Proceedings of the 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Vancouver, BC, Canada, 17–24 June 2023; pp. 2818–2829. [Google Scholar] [CrossRef]
  135. Tooth, J.M.; Tuptuk, N.; Watson, J.D.M. A Systematic Survey of the Gemini Principles for Digital Twin Ontologies, 2024. Version Number: 1. [CrossRef]
  136. D’Amico, R.D.; Sarkar, A.; Karray, M.H.; Addepalli, S.; Erkoyuncu, J.A. Knowledge transfer in Digital Twins: The methodology to develop Cognitive Digital Twins. CIRP J. Manuf. Sci. Technol. 2024, 52, 366–385. [Google Scholar] [CrossRef]
Figure 1. Process diagram pt.1: automatic ontology term and RDF triples extraction from cultural heritage reports.
Figure 1. Process diagram pt.1: automatic ontology term and RDF triples extraction from cultural heritage reports.
Applsci 15 10061 g001
Figure 2. Classes and properties used by ARTEMIS to model heritage science information.
Figure 2. Classes and properties used by ARTEMIS to model heritage science information.
Applsci 15 10061 g002
Figure 3. Graph visualisation of a scientific analysis of a cultural heritage asset directly obtained from the PDF through our AI-assisted pipeline.
Figure 3. Graph visualisation of a scientific analysis of a cultural heritage asset directly obtained from the PDF through our AI-assisted pipeline.
Applsci 15 10061 g003
Figure 4. Process diagram pt.2: querying the KG in natural language using CompGCN embeddings.
Figure 4. Process diagram pt.2: querying the KG in natural language using CompGCN embeddings.
Applsci 15 10061 g004
Figure 5. Schematic of CompGCN message passing in cultural heritage knowledge graphs. Example: Painting aggregates messages from Artist and Exhibition. Dashed red arrows indicate composed messages being aggregated into the entity embeddings.
Figure 5. Schematic of CompGCN message passing in cultural heritage knowledge graphs. Example: Painting aggregates messages from Artist and Exhibition. Dashed red arrows indicate composed messages being aggregated into the entity embeddings.
Applsci 15 10061 g005
Figure 6. Diagram of the querying process using CompGCN embeddings and similarity computation.
Figure 6. Diagram of the querying process using CompGCN embeddings and similarity computation.
Applsci 15 10061 g006
Figure 7. Snapshot of the functional chat-UI demonstrating a user prompt and the query results in narrative format.
Figure 7. Snapshot of the functional chat-UI demonstrating a user prompt and the query results in narrative format.
Applsci 15 10061 g007
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Felicetti, A.; Himmiche, A.; Somenzi, M. Knowledge Graphs and Artificial Intelligence for the Implementation of Cognitive Heritage Digital Twins. Appl. Sci. 2025, 15, 10061. https://doi.org/10.3390/app151810061

AMA Style

Felicetti A, Himmiche A, Somenzi M. Knowledge Graphs and Artificial Intelligence for the Implementation of Cognitive Heritage Digital Twins. Applied Sciences. 2025; 15(18):10061. https://doi.org/10.3390/app151810061

Chicago/Turabian Style

Felicetti, Achille, Aida Himmiche, and Miriana Somenzi. 2025. "Knowledge Graphs and Artificial Intelligence for the Implementation of Cognitive Heritage Digital Twins" Applied Sciences 15, no. 18: 10061. https://doi.org/10.3390/app151810061

APA Style

Felicetti, A., Himmiche, A., & Somenzi, M. (2025). Knowledge Graphs and Artificial Intelligence for the Implementation of Cognitive Heritage Digital Twins. Applied Sciences, 15(18), 10061. https://doi.org/10.3390/app151810061

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop