ARIA: An AI-Supported Adaptive Augmented Reality Framework for Cultural Heritage
Abstract
1. Introduction
- Articulating an AR-first, modular platform architecture that explicitly links immersive delivery, sensing/monitoring, DAM, and adaptive AI components, which are often treated separately in prior work;
- Formalizing an affect–context fusion loop that specifies how affective proxies participate in real-time narrative and interface adaptation, rather than remaining at the level of post-hoc analytics;
- Embedding RAG-based personalization within the AR adaptation pipeline and grounding generated narratives in curator-approved institutional knowledge, in contrast to standalone LLM-based museum assistants;
- Operationalizing ethical and inclusivity principles through concrete governance mechanisms (privacy, consent, XAI, bias mitigation, authenticity checks), rather than addressing them as abstract guidelines;
- Supporting feasibility and evaluability through a component-to-metric validation mapping, enabling structured walkthrough-based assessment instead of relying solely on illustrative scenarios.
2. State of the Art and Related Work
2.1. Evidence from Surveys and Literature Mappings
2.2. Existing Frameworks for AR in Cultural Heritage and Positioning of ARIA
2.3. Architectural Frameworks for AR and XR in Cultural Heritage
2.4. Usability, Interaction, and Fragmentation in Extended Reality
2.5. Limitations in Adaptation, Personalization, and Affective Computing
2.6. Ethical and Human-Centered Design Challenges
2.7. Synthesized Research Gap
3. The Augmented Reality for Interpreting Artefacts (ARIA) Framework
3.1. Conceptual Overview and Modular Architecture
- Immersive XR Module: This module integrates primarily AR-based visualization, with optional MR/VR extensions depending on institutional needs. A core focus is the generation of high-fidelity 3D assets using polygonal mesh pipelines, photogrammetry, and high-resolution multimodal digitization [39]. Standardized formats enhance cross-museum interoperability and long-term digital preservation, addressing gaps highlighted in current XR deployments.
- Adaptive AI Module: The AI module is the platform’s decision-making core, enabling context-aware, affect-responsive, and hyper-personalized content delivery. It integrates multimodal analytics (interaction patterns, sensor-based observations, and emotional cues) with advanced recommendation engines and Retrieval-Augmented Generation (RAG) models. Continuous model refinement can be supported through a shared learning pool, aligning with state-of-the-art adaptive systems in CH [40].
- Data Collection and Monitoring Module: This module aggregates real-time data from IoT sensors, computer-vision tools, and interaction logs, enabling precise visitor flow analysis, engagement mapping, and affective state estimation. Real-time data pipelines are central to delivering low-latency adaptive content, reflecting recent developments in responsive museum analytics [41]. All sensing modalities are considered optional and configurable, subject to explicit consent and institutional policy constraints.
- Digital Asset Management (DAM) Module: A centralized DAM repository supports FAIR-compliant management of digital artifacts, enabling advanced semantic enrichment using AI-based metadata extraction [42]. This module addresses the widespread fragmentation of digital assets across GLAM institutions.
- Cultural Intelligence Dashboard: This dashboard consolidates multimodal analytics and KPI monitoring into an operational interface for curators and administrators. It supports evidence-informed curatorial reflection, exhibit planning, and adaptive interpretation strategies. Cultural decision-support environments have been identified as a critical missing element in digital museology workflows [43].
3.2. Foundational Methodology and Design Principles
- Privacy-by-Design and GDPR Compliance: All data processing pipelines incorporate minimization, anonymization, and local processing strategies.
- Explainable AI (XAI): Transparency layers inform visitors how recommendations, reconstructions, or adaptive narratives were produced [48].
- Bias Mitigation: AI models are trained on diverse datasets, with human-expert review loops to prevent cultural distortion or stereotyping [49].
3.3. Illustrative Use Case Scenario
3.4. End-to-End Adaptive Loop and Data-Flow Sequencing
- Onboarding, consent, and initialization. The visitor initiates the AR experience through a mobile or wearable AR interface. During onboarding, consent preferences are explicitly captured (e.g., location tracking, interaction logging, optional affect proxies), configuring which sensing and personalization modules are activated. This step corresponds to the Ethical Governance and Immersive XR modules in Figure 1 and sets the boundary conditions for the entire adaptive process.
- Contextual sensing and proxy acquisition. As the visitor navigates the physical space, the Data Collection and Monitoring module continuously gathers multimodal contextual signals, including spatial position, dwell time, navigation patterns, and interaction events. Where permitted, lightweight affective proxies (e.g., hesitation, pacing, repeated interactions) are inferred without relying on intrusive biometric measurements. This stage operationalizes the Context-Aware and Affective Integration pillar (Figure 3).
- Fusion and user model update. Incoming contextual and affective signals are fused and used to incrementally update the visitor’s dynamic user model. The model captures evolving interests, engagement levels, and accessibility needs, forming the decision state that drives personalization. This process bridges the sensing layer with the Adaptive AI module in Figure 1.
- Grounded retrieval and adaptive reasoning. Based on the updated user model and situational context, the Adaptive AI module triggers a Retrieval-Augmented Generation (RAG) process. Curator-approved knowledge sources and DAM-managed assets are queried to retrieve relevant factual content, which constrains and grounds any generated narrative elements. This step directly maps to the AI-Driven Personalization and Adaptive Storytelling pillar.
- Narrative and interface adaptation. Retrieved and generated content is translated into adaptive AR narratives and interface configurations. Narrative depth, modality (text, audio, visual overlay), pacing, and interaction complexity are adjusted in real time to match the visitor’s engagement state and contextual conditions. This stage connects the AI reasoning outputs to the Immersive XR Module, producing the experiential outcomes illustrated in Figure 2.
- Experience logging and feedback aggregation. Interaction events, adaptation decisions, and high-level engagement indicators are logged in anonymized form. These data streams feed the Cultural Intelligence Dashboard, enabling curators and administrators to observe aggregate patterns, evaluate interpretive strategies, and identify content or interaction bottlenecks.
- Curatorial reflection and system refinement. Insights derived from the dashboard inform curatorial decision-making, content updates, and system tuning. This closes the adaptive loop by linking visitor-facing experiences back to institutional governance and long-term experience design.
4. ARIA’s Innovative Technological Pillars for Adaptive AR
4.1. Real-Time Context-Aware and Affective AR Integration
- Multi-Sensor Data Fusion: The ARIA framework incorporates a dense layer of IoT devices, environmental sensors, and tracking systems to continuously monitor visitor location, movement, crowd density, and ambient conditions. Similar IoT-based monitoring solutions have proven effective for visitor analytics and flow optimization in museum contexts [41]. By aggregating these heterogeneous data streams, the platform can construct a fine-grained model of the visitor’s situational context (e.g., congestion, dwell time, preferred paths), overcoming the limited, single-source triggers typical of current AR installations [10].
- Emotion Recognition and Affective Computing: A second component of this pillar is the integration of emotion recognition and affective computing to assess engagement, confusion, or fatigue in real time. Prior work has demonstrated the potential of affect-aware systems in education and cultural heritage to improve adaptive support and user satisfaction [13,30]. In the ARIA, multimodal cues such as facial expressions, body posture, dwell patterns, and interaction rhythms can be used to infer affective and cognitive states, under strict ethical and privacy constraints (see Section 5). This directly responds to the identified gap whereby affective sensing in CH remains at low TRLs and largely confined to experimental prototypes [30].
- Dynamic AR Content Adaptation: The fusion of contextual and affective signals enables dynamic modulation of AR content. For instance, if a visitor exhibits signs of confusion or reduced engagement, the system can simplify explanations, provide additional clarifications, or redirect them toward complementary exhibits. Conversely, highly engaged visitors may receive deeper, more expert-level narratives or alternative interpretive perspectives. This fine-grained adaptivity addresses the dominance of static, one-size-fits-all AR experiences noted in recent reviews [12,20] and aligns with emerging paradigms of responsive, data-driven museum interpretation [43].
4.2. AI-Driven Personalization and Adaptive Storytelling
- RAG-Based Conversational Agents: RAG-based agents combine Large Language Models (LLMs) with curated cultural heritage knowledge bases and institution-specific datasets [37,40]. Instead of generic chatbots that offer static or loosely curated information [31], these agents can retrieve factual, institution-approved content and generate explanations, comparisons, and narratives aligned with the visitor’s current location, interaction history, and preferences. This setup reduces the risk of hallucinated information by grounding generative output in verified collections data and expert-curated metadata [42,50].
- Sophisticated User Modeling: The ARIA maintains continuously evolving user models that integrate explicit preferences (e.g., chosen themes, languages), implicit behaviors (e.g., dwell time, navigation patterns), and inferred interests (e.g., repeated focus on specific cultural periods or media types). Such models build on recent work in adaptive personalization for CH [40], extending it with affective and contextual features harvested through the data collection and monitoring module [41]. This enables adaptive pacing, difficulty scaling, and content depth control, aiming to prevent both cognitive overload and under-stimulation [21,46].
- Semantic Relationship Modeling and Emergent Narratives: To support rich, non-linear narrative paths, the ARIA employs semantic relationship modeling across digital objects, narratives, and contextual entities. Knowledge-graph-based approaches and AI-based metadata enrichment [42,43] are used to define conceptual, temporal, and spatial relationships between artifacts. As visitors navigate the physical and digital space, the system can dynamically surface connections (e.g., shared motifs, techniques, historical events) and generate emergent storylines tailored to their interests, addressing calls for more interpretatively sophisticated XR experiences in CH [27,28,39]. This semantic modeling architecture is novel in that it allows narrative pathways to emerge dynamically from visitor behavior and emotional responses, transforming AR storytelling from a static linear sequence into a more flexible, generative cultural knowledge experience.
4.3. High-Fidelity XR Content Creation and Digital Asset Management
- High-Fidelity Digital Twins: The ARIA adopts state-of-the-art 3D recording and photogrammetric workflows to create high-fidelity digital twins of artifacts, architectural elements, and entire exhibition spaces [39]. Dense multi-view imaging, laser scanning, and mesh optimization pipelines support the production of detailed polygonal models suitable for AR/MR deployment while preserving conservation-grade accuracy. Such models underpin both in-situ AR overlays and immersive VR reconstructions, improving the realism and credibility of digital experiences [20,39].
- AI-Based Digital Asset Management: To overcome fragmented DAM ecosystems, the ARIA integrates an AI-enhanced Digital Asset Management module that performs automated semantic enrichment, entity extraction, and relationship discovery across collections [10,42]. This reduces the manual burden on curators, improves cross-collection discoverability, and supports FAIR-compliant data practices [38,42]. The enriched metadata feed directly into the AI personalization and storytelling components, creating a virtuous cycle between content management and visitor-facing adaptivity. The DAM module extends existing systems by incorporating AI-based metadata enrichment that continuously aligns digital assets with evolving user models and emotional states. This tight coupling between semantic asset management and adaptive interaction has not been systematically documented within an AR-first cultural heritage framework [40,43].
4.4. Multimodal Interaction and Seamless Phygital Integration
- Seamless Transitions and Spatial-Temporal Synchronization: A key objective is to maintain a unified sense of presence across physical and digital layers. Existing MR systems often suffer from misaligned overlays, tracking drift, or abrupt transitions, which disrupt immersion and increase cognitive load [9,21]. The ARIA architecture therefore emphasizes robust spatial registration, dynamic mapping, and temporal synchronization of content streams. Building on empirical insights from XR cultural heritage evaluation [2,20], the platform seeks to minimize visual discontinuities and to coordinate AR content with visitor movement and gaze.
- Multimodal Interaction Channels: To support heterogeneous user needs and abilities, the ARIA enables multimodal interaction, including gaze-based selection, midair gestures, touch input, voice commands, and haptic feedback where available [16,46]. This design responds to recent work in inclusive interface design and adaptive accessibility [34,46], recognizing that no single interaction paradigm is universally suitable in museum contexts. Latency-sensitive operations (e.g., gesture recognition, spatial mapping) can be offloaded to edge-computing nodes to reduce lag and motion sickness [9,41].
- Ergonomics and Cognitive Load Management: Finally, the ARIA explicitly integrates ergonomic and cognitive design considerations, drawing on established HCI and VR design guidelines [21,44,46]. Interface layouts, information density, and interaction complexity are tuned dynamically according to user profiles and real-time signals (e.g., signs of fatigue or overload), complementing the affective adaptivity described above. This pillar aims to counteract the usability issues and discomfort that often limit the adoption and long-term use of XR in CH settings [16,20,21,46].
4.5. Implementation Challenges and Technical Feasibility
4.6. Minimum Viable Deployment, Resources, and Feasibility Constraints
- one curator or domain expert responsible for content validation and narrative constraints;
- one XR developer for AR interface and interaction logic; and
- one AI or software engineer responsible for integration of retrieval, recommendation, and analytics services. Advanced affective computing or multimodal sensing components may be introduced incrementally as institutional expertise and resources mature.
5. Ethics, Human-Centered Design, and Formative Validation
5.1. Ethical Governance and Privacy-Conscious Design
- Data Minimization and GDPR Compliance: Compliance with the General Data Protection Regulation (GDPR) is ensured through strict data minimization, purpose limitation, and secure processing. Sensitive signals such as emotional cues are anonymized, encrypted, and processed locally whenever possible to reduce systemic risk, aligning with current best practices in privacy-preserving AI [32,53].
- Consent and User Autonomy: Visitors retain full control over the use of their data through explicit, granular opt-in mechanisms for camera access, emotion recognition, and sensor logging. Users may revoke consent at any time, supporting autonomy and informed participation [32]. These mechanisms are crucial for trust-building in GLAM environments, where technological interventions intersect with cultural sensitivities.
- Transparency via Explainable AI (XAI): Explainable AI techniques enhance transparency by enabling curators and visitors to understand how recommendations, adaptive narratives, and XR reconstructions are produced. In line with recent XAI developments [48], ARIA integrates traceability layers and AR-based explanatory overlays that reveal data sources, model reasoning, and curatorial decision pathways. This is essential to maintaining institutional trust and preventing opaque or unintentionally biased interpretations.
- Cultural Integrity and Bias Mitigation: AI-driven personalization risks introducing cultural biases or historically inaccurate “hallucinated” content. ARIA addresses these risks through: (i) mixed-initiative, template-based narrative generation to constrain AI outputs; (ii) expert-in-the-loop validation for digital reconstructions; and (iii) bias mitigation processes grounded in diverse training data and regular fairness audits [49,50]. Recent studies highlight the presence of cultural bias in museum datasets and knowledge graphs [54], underscoring the importance of these measures.
5.2. Human-Centered Design, Inclusivity, and Co-Creation
- Participatory and Co-Creation Methodology: The framework promotes participatory design involving curators, educators, accessibility experts, and community stakeholders from early ideation to deployment. Co-creation facilitates shared authorship of interpretative narratives and strengthens cultural authenticity [45]. Such interdisciplinary collaboration is particularly effective in contextualizing AI-driven experiences within culturally sensitive domains.
- Universal Design and Inclusive Interaction: ARIA operationalizes Universal Design principles to support visitors with diverse sensory, cognitive, linguistic, and motor needs. Strategies include adaptive multimodal interfaces, multilingual content, mobility-friendly navigation, and accessible XR interaction modalities. This approach is consistent with research on inclusive XR design [46,55].
- Scenario-Based Design and Requirements Elicitation: Scenario-Based Design (SBD) is employed to capture experiential requirements and explore use cases for visitors with different accessibility profiles [44]. SBD supports the design of elastic interfaces, reduces cognitive load, and ensures that interaction models remain intuitive in mixed-reality contexts.
5.3. Validation Strategy and Preliminary Formative Results
- Security and Privacy Compliance: architectural audits, penetration testing, and GDPR compliance checks ensure robust data protection and system resilience [53].
5.3.1. Implementation Constraints and Trade-Offs
5.3.2. Preliminary Formative Validation Results
- Expert walkthrough evaluation: A structured expert walkthrough was performed to evaluate the architectural completeness and controllability of the adaptive loop. The walkthrough involved domain experts from cultural heritage and human–computer interaction (e.g., curators, digital heritage researchers, and HCI practitioners; ). Experts were guided through the end-to-end adaptation sequence (onboarding, sensing, user model update, adaptive reasoning, AR delivery, and curatorial feedback) and asked to assess feasibility, clarity of responsibilities, and alignment with GLAM institutional practices. Feedback confirmed that the modular decomposition and sequencing reflect realistic museum workflows and improve transparency compared to monolithic or ad hoc AR system designs. The walkthrough explicitly leveraged the illustrative use-case scenarios and end-to-end adaptation sequence to guide expert assessment of system behavior and decision traceability.
- Scenario-based consistency validation: In addition, a scenario-based validation was conducted using representative visitor profiles (e.g., casual visitor, expert visitor, accessibility-focused visitor). For each scenario, the adaptation logic was traced step by step to verify that experiential outcomes illustrated in the use-case view can be derived from the proposed architectural modules and technological pillars without introducing undocumented assumptions. This analysis demonstrated internal consistency between the architecture (Figure 1), the technological pillars (Figure 3), and the visitor outcomes (Figure 2), while also confirming compliance with governance constraints such as consent configuration and curator oversight.
- Technical feasibility reasoning: Finally, a feasibility check was performed at an analytical level to assess whether the adaptive loop can operate under realistic performance constraints. The pipeline was decomposed into latency-sensitive operations (e.g., interaction tracking, spatial alignment) and compute-intensive processes (e.g., retrieval and narrative generation). This reasoning supports the plausibility of hybrid deployment strategies combining on-device, edge, and cloud components, indicating that the framework does not rely on unrealistic computational assumptions.
6. Conclusions and Future Research Directions
6.1. Summary of Contributions and Implications
- Theoretical implications: From a theoretical perspective, ARIA advances the state of the art in several ways. First, it introduces a unifying AR-first architectural abstraction for adaptive cultural interpretation, clarifying the role of Augmented Reality as a primary medium for in-situ meaning-making rather than as a secondary component within broader XR ecosystems. Second, the framework explicitly defines an end-to-end affect- and context-informed adaptation loop, operationalizing how sensing, user modeling, adaptive reasoning, and AR delivery interact in real time—an aspect that remains implicit or absent in most prior frameworks. Third, ARIA contributes to responsible AI discourse in cultural heritage by translating ethical and governance principles (e.g., privacy, transparency, accountability, curatorial authority) into concrete architectural mechanisms, rather than treating them as abstract guidelines. Finally, the framework establishes a structured bridge between AI-grounded narrative adaptation (via Retrieval-Augmented Generation) and museological authority, demonstrating how generative techniques can be constrained, explainable, and curator-controlled within interpretive systems.
- Practical implications: From a practical standpoint, ARIA provides actionable guidance for GLAM institutions, system designers, and practitioners. The framework functions as a blueprint for incremental adoption, allowing institutions to deploy selected modules (e.g., context-aware AR or DAM integration) without committing to full-scale system transformation. It clarifies operational roles and control points for curators and heritage professionals, including approval workflows, narrative constraints, and oversight of adaptive behavior. ARIA also offers concrete guidance on integrating Digital Asset Management and metadata infrastructures with adaptive AR delivery, supporting interoperability and reuse across exhibitions and institutions. Finally, by explicitly mapping framework components to validation metrics, the framework supports evaluation-ready deployments, facilitating feasibility assessment, pilot studies, and evidence-informed procurement decisions.
6.2. Limitations of the Present Study
6.3. Future Research Directions
- Prototype-based and pilot validation: Developing partial implementations (e.g., RAG-driven narrative adaptation without affect sensing) to empirically test feasibility, latency, and user acceptance in controlled museum settings.
- Robust affect proxies for cultural environments: Investigating culturally sensitive, privacy-preserving affect indicators suitable for real-world GLAM contexts and systematically documenting their failure modes.
- Longitudinal impact studies: Examining how adaptive AR influences learning outcomes, interpretive depth, and visitor memory over extended periods, beyond immediate engagement metrics.
- Cross-institution interoperability: Exploring shared metadata standards, semantic alignment strategies, and collaborative DAM infrastructures to support adaptive AR across institutional boundaries.
- Governance and organizational adoption: Studying how curatorial practices, staffing models, and institutional policies shape the adoption of adaptive AI systems in museums and heritage sites.
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Acknowledgments
Conflicts of Interest
Abbreviations
| ARIA | Augmented Reality for Interpreting Artefacts |
| AI | Artificial Intelligence |
| AR | Augmented Reality |
| CH | Cultural Heritage |
| DAM | Digital Asset Management |
| FATE | Fairness, Accountability, Transparency, and Ethics |
| FAIR | Findable, Accessible, Interoperable, Reusable |
| GDPR | General Data Protection Regulation |
| GLAM | Galleries, Libraries, Archives, and Museums |
| HCD | Human-Centered Design |
| HCI | Human–Computer Interaction |
| HMD | Head-Mounted Display |
| IoT | Internet of Things |
| KPI | Key Performance Indicator |
| LLM | Large Language Model |
| MR | Mixed Reality |
| RAG | Retrieval-Augmented Generation |
| SBD | Scenario-Based Design |
| SotA | State of the Art |
| TRL | Technology Readiness Level |
| UI | User Interface |
| UX | User Experience |
| VR | Virtual Reality |
| XR | Extended Reality |
References
- Bekele, M.K.; Pierdicca, R.; Frontoni, E.; Malinverni, E.S.; Gain, J. A Survey of Augmented Reality Applications in Cultural Heritage. Comput. Graph. 2018, 77, 1–20. [Google Scholar]
- Tromp, J.; Schofield, D.; Raeisian Parvari, P.; Poyade, M.; Eaglesham, C.; Torres, J.C.; Johnson, T.; Jürivete, T.; Lauer, N.; Reyes-Lecuona, A.; et al. Designing and Evaluating XR Cultural Heritage Applications Through HCI Methods. Appl. Sci. 2025, 15, 7973. [Google Scholar] [CrossRef]
- Sylaiou, S.; Mania, K.; Karoulis, A.; White, M. Exploring the Relationship Between Presence and Enjoyment in a Virtual Museum. Int. J. Hum.–Comput. Stud. 2010, 68, 243–253. [Google Scholar] [CrossRef]
- Pujol, L.; Champion, E. Evaluating Presence in Cultural Heritage Projects. Int. J. Herit. Stud. 2012, 18, 83–102. [Google Scholar] [CrossRef]
- Konstantakis, M.; Aliprantis, J.; Teneketzis, A.; Caridakis, G. Understanding User Experience Aspects in Cultural Heritage Interaction. In Proceedings of the 22nd Pan-Hellenic Conference on Informatics, Athens, Greece, 29 November–1 December 2018; ACM: New York, NY, USA, 2018; pp. 267–271. [Google Scholar]
- Komianos, V. Immersive applications in museums: An analysis of the use of XR technologies and the provided functionality based on systematic literature review. JOIV Int. J. Inform. Vis. 2022, 6, 60–73. [Google Scholar] [CrossRef]
- Zhou, Y.; Chen, J.; Wang, M. A meta-analytic review on incorporating virtual and augmented reality in museum learning. Educ. Res. Rev. 2022, 36, 100454. [Google Scholar] [CrossRef]
- Innocente, C.; Ulrich, L.; Moos, S.; Vezzetti, E. A framework study on the use of immersive XR technologies in the cultural heritage domain. J. Cult. Herit. 2023, 62, 268–283. [Google Scholar] [CrossRef]
- Kobeisse, S. Hands on the Past: Towards a Conceptual Framework for Developing and Evaluating Tangible AR Interfaces for Historical Artefacts. In Proceedings of the 35th Australian Computer-Human Interaction Conference, Wellington, New Zealand, 2–6 December 2023. [Google Scholar]
- Boboc, R.G.; Băutu, E.; Gîrbacia, F.; Popovici, N.; Popovici, D.M. Augmented reality in cultural heritage: An overview of the last decade of applications. Appl. Sci. 2022, 12, 9859. [Google Scholar] [CrossRef]
- Breque, M.; de Nul, L.; Petridis, A. Industry 5.0: Towards a Sustainable, Human-Centric Industrial Model; European Commission Report; European Commission: Brussels, Belgium, 2021. [Google Scholar]
- Mortara, M.; Catalano, C.E.; Bellotti, F.; Fiucci, G.; Houry-Panchetti, M.; Petridis, P. Learning Cultural Heritage Through Serious Games: A Review. J. Cult. Herit. 2014, 15, 318–329. [Google Scholar] [CrossRef]
- Yun, H.J.; Cho, J. Affective domain studies of K-12 computing education: A systematic review from a perspective on affective objectives. J. Comput. Educ. 2022, 9, 477–514. [Google Scholar] [CrossRef]
- Konstantakis, M.; Aliprantis, J.; Michalakis, K.; Caridakis, G. Recommending user experiences based on extracted cultural personas for mobile applications-REPEAT methodology. In Proceedings of the 20th International Conference on Human-Computer Interaction with Mobile Devices and Services (MobileHCI’18), Barcelona, Spain, 3–6 September 2018. [Google Scholar]
- Konstantakis, M.; Kalatha, E.; Caridakis, G. Cultural heritage, serious games and user personas based on gardner’s theory of multiple intelligences:“The Stolen Painting” game. In International Conference on Games and Learning Alliance; Springer International Publishing: Cham, Switzerland, 2019; pp. 490–500. [Google Scholar]
- Aliprantis, J.; Kalatha, E.; Konstantakis, M.; Michalakis, K.; Caridakis, G. Linked Open Data as Universal Markers for Mobile Augmented Reality Applications in Cultural Heritage. In Digital Cultural Heritage: Final Conference of the Marie Skłodowska-Curie Initial Training Network for Digital Cultural Heritage, ITN-DCH 2017, Olimje, Slovenia, 23–25 May 2017; Revised Selected Papers; Springer International Publishing: Cham, Switzerland, 2018; pp. 79–90. [Google Scholar]
- Floridi, L.; Cowls, J. A Unified Framework of Five Principles for AI in Society. Harv. Data Sci. Rev. 2019, 1, 1–15. [Google Scholar] [CrossRef]
- Konstantakis, M.; Caridakis, G. Adding Culture to UX: UX Research Methodologies and Applications in Cultural Heritage. J. Comput. Cult. Herit. 2020, 13, 1–17. [Google Scholar] [CrossRef]
- Ordoumpozanis, K.; Konstantakis, M.; Zoi, S.; Caridakis, G. Generative AI: A Systematic Review of Related Interfaces and Interactions. In Proceedings of the 3rd International Conference of the ACM Greek SIGCHI Chapter, Syros, Greece, 24–26 September 2025; ACM: New York, NY, USA, 2025; pp. 39–47. [Google Scholar]
- Caarls, J.; Jonker, P.; Kolstee, Y.; Rotteveel, J.; van Eck, W. Augmented Reality for Art, Design and Cultural Heritage—System Design and Evaluation. EURASIP J. Image Video Process. 2009, 2009, 716160. [Google Scholar] [CrossRef]
- Jerald, J. The VR Book: Human-Centered Design for Virtual Reality; Morgan & Claypool: San Rafael, CA, USA, 2016. [Google Scholar]
- Zarantonello, L.; Schmitt, B.H. Experiential AR/VR: A consumer and service framework and research agenda. J. Serv. Manag. 2023, 34, 34–55. [Google Scholar] [CrossRef]
- Bozzelli, G.; Raia, A.; Ricciardi, S.; De Nino, M.; Barile, N.; Perrella, M.; Tramontano, M.; Pagano, A.; Palombini, A. An integrated VR/AR framework for user-centric interactive experience of cultural heritage: The ArkaeVision project. Digit. Appl. Archaeol. Cult. Herit. 2019, 15, e00124. [Google Scholar] [CrossRef]
- Hammady, R.; Ma, M.; Al-Kalha, Z.; Strathearn, C. A framework for constructing and evaluating the role of MR as a holographic virtual guide in museums. Virtual Real. 2021, 25, 895–918. [Google Scholar] [CrossRef]
- Spadoni, E.; Carulli, M.; Bordegoni, M. A Conceptual Framework to Support a New Collaborative Design Process for Immersive Technology Integration in Museum Exhibitions. In International Conference on Extended Reality; Springer Nature: Cham, Switzerland, 2023; pp. 160–178. [Google Scholar]
- Li, J.; Wider, W.; Ochiai, Y.; Fauzi, M.A. A bibliometric analysis of immersive technology in museum exhibitions: Exploring user experience. Front. Virtual Real. 2023, 4, 1240562. [Google Scholar] [CrossRef]
- Champion, E. Critical Gaming: Interactive History and Virtual Heritage; Routledge: London, UK, 2016. [Google Scholar]
- Balbi, B.; Marasco, A. Co-Designing Cultural Heritage Experiences for All with Virtual Reality. Um. Digit. 2021, 11, 71–94. [Google Scholar]
- Bellotti, F.; Berta, R.; De Gloria, A.; D’Ursi, A.; Fiore, V. A Serious Game Model for Cultural Heritage. J. Comput. Cult. Herit. (JOCCH) 2013, 5, 1–27. [Google Scholar] [CrossRef]
- Hu, H.; Wan, Y.; Tang, K.Y.; Li, Q.; Wang, X. Affective-Computing-Driven Personalized Display of Cultural Information for Commercial Heritage Architecture. Appl. Sci. 2025, 15, 3459. [Google Scholar] [CrossRef]
- De Mello, J.C.; Secci, G.P.N.; Ribeiro, P.H. Robotics and AI in museums—The future of the present. Robot. Syst. Appl. 2024, 4, 44–58. [Google Scholar] [CrossRef]
- Tavani, H. Ethics and Technology: Controversies, Questions, and Strategies for Ethical Computing; Wiley: New York, NY, USA, 2016. [Google Scholar]
- Floridi, L.; Cowls, J. A unified framework of five principles for AI in society. In Machine Learning and the City: Applications in Architecture and Urban Design; John Wiley & Sons Ltd.: Hoboken, NJ, USA, 2022; Volume 34, pp. 535–545. [Google Scholar]
- Atata, R.; Odedeyi, T. Designing Inclusive Interfaces: Enhancing User Experience for People with Disabilities Through Adaptive UI Accessibility Principles. Int. J. Sci. Res. Arch. 2025, 16, 1072–1096. [Google Scholar] [CrossRef]
- Tiribelli, S.; Pansoni, S.; Frontoni, E.; Giovanola, B. Ethics of artificial intelligence for cultural heritage: Opportunities and challenges. IEEE Trans. Technol. Soc. 2024, 5, 293–305. [Google Scholar] [CrossRef]
- Akman, A.Z.; Erdirençelebi, M. A human-centered digital transformation: A bibliometric analysis of society 5.0 and industry 5.0. Istanb. Manag. J. 2024, 2, 1361480. [Google Scholar] [CrossRef]
- Wu, H.-C.; Guo, W. AI Adoption in the Creative Sector: Practices, Perceptions, and Prospects Among Arts and Culture Organizations and Administrators. In The Art of Digital Governance: Navigating Platforms and AI Revolution; Springer Nature: Cham, Switzerland, 2025; pp. 407–431. [Google Scholar]
- European Commission. European Collaborative Cloud for Cultural Heritage (ECCCH): Strategic Orientation; Publications Office of the EU: Luxembourg, 2022. [Google Scholar]
- Konstantakis, M.; Trichopoulos, G.; Aliprantis, J.; Gavogiannis, N.; Karagianni, A.; Parthenios, P.; Caridakis, G. An Improved Approach for Generating Digital Twins of Cultural Spaces Through the Integration of Photogrammetry and Laser Scanning Technologies. Digital 2024, 4, 215–231. [Google Scholar] [CrossRef]
- Alexandridis, G.; Chrysanthi, A.; Tsekouras, G.E.; Caridakis, G. Personalized and content adaptive cultural heritage path recommendation: An application to the Gournia and Çatalhöyük archaeological site. User Model.-User-Adapt. Interact. 2019, 29, 201–238. [Google Scholar]
- Ivanov, R.; Velkova, V. Analyzing Visitor Behavior to Enhance Personalized Experiences in Smart Museums: A Systematic Literature Review. Computers 2025, 14, 191. [Google Scholar] [CrossRef]
- Baek, S.; Hwang, H.; Park, C.W.; Kim, H.K.; Lee, J.H. Development of an Artificial Intelligence-based Platform for the Analysis and Utilization of Cultural Heritage Data. ISPRS Ann. Photogramm. Remote Sens. Spat. Inf. Sci. 2025, X-M-2-2025, 39–46. [Google Scholar] [CrossRef]
- Münster, S.; Utescher, R.; Ulutaş Aydogan, S. Digital Topics on Cultural Heritage Investigated: How Can Data-Driven and Data-Guided Methods Support to Identify Current Topics and Trends in Digital Heritage? Built Herit. 2021, 5, 25. [Google Scholar] [CrossRef] [PubMed]
- Pisoni, G.; Díaz-Rodríguez, N.; Gijlers, H.; Tonolli, L. Human-Centered Artificial Intelligence for Designing Accessible Cultural Heritage. Appl. Sci. 2021, 11, 870. [Google Scholar] [CrossRef]
- Sanders, E.; Stappers, P. Co-Creation and the New Landscapes of Design. CoDesign 2008, 4, 5–18. [Google Scholar] [CrossRef]
- Fox, D.; Thornton, I.G. White Paper-The IEEE Global Initiative on Ethics of Extended Reality (XR) Report–Extended Reality (XR) Ethics and Diversity, Inclusion, and Accessibility. In The IEEE Global Initiative on Ethics of Extended Reality (XR) Report–Extended Reality (XR) Ethics and Diversity, Inclusion, and Accessibility; IEEE: Piscataway, NJ, USA, 2022; pp. 1–25. [Google Scholar]
- Allou, E. A Universal Ethical Framework for AI Systems Based on the Pure Time Theory; Springer: Berlin/Heidelberg, Germany, 2025. [Google Scholar]
- Belle, V.; Papantonis, I. Principles and Practice of Explainable Machine Learning. Front. Big Data 2021, 4, 688969. [Google Scholar] [CrossRef]
- Kumar, A.; Aelgani, V.; Vohra, R.; Gupta, S.K.; Bhagawati, M.; Paul, S.; Suri, J.S. Artificial Intelligence Bias in Medical System Designs: A Systematic Review. Multimed. Tools Appl. 2024, 83, 18005–18057. [Google Scholar]
- Abdelmonem, M.G. Architectural and Urban Heritage in the Digital Age: Dilemmas of Authenticity, Originality and Reproduction. Int. J. Archit. Res. ArchNet-IJAR 2017, 11, 5–15. [Google Scholar]
- Martella, C.; Miraglia, A.; Frost, J.; Cattani, M.; Van Steen, M. Visualizing, clustering, and predicting the behavior of museum visitors. Pervasive Mob. Comput. 2017, 38, 430–443. [Google Scholar] [CrossRef]
- Wang, Z.; Qiu, J. Research on the construction of revolutionary cultural heritage knowledge graph and the application of digital cultural innovation. Acad. J. Humanit. Soc. Sci. 2024, 7, 143–148. [Google Scholar] [CrossRef]
- Chang, J.M.; Zhuang, D.; Samaraweera, G.D. Privacy-Preserving Machine Learning; Simon and Schuster: New York, NY, USA, 2021. [Google Scholar]
- Bobasheva, A.; Gandon, F.; Precioso, F. Learning and Reasoning for Cultural Metadata Quality: Coupling Symbolic AI and Machine Learning over a Semantic Web Knowledge Graph to Support Museum Curators in Improving the Quality of Cultural Metadata and Information Retrieval. J. Comput. Cult. Herit. (JOCCH) 2022, 15, 1–23. [Google Scholar] [CrossRef]
- Saviano, M.; Malakuczi, V.; Imbesi, L. Toward Accessible Mixed Reality in the Museum: Usability Principles for Disadvantaged Visitors and a Feasibility Evaluation for a New MR System. Int. J. Des. Soc. 2024, 18, 65. [Google Scholar] [CrossRef]
- Qian, W.; Coutinho, R.W. On the design of edge-assisted mobile IoT augmented and mixed reality applications. In Proceedings of the 17th ACM Symposium on QoS and Security for Wireless and Mobile Networks, Alicante, Spain, 22 November 2021; pp. 131–136. [Google Scholar]
- Lewis, P.; Perez, E.; Piktus, A.; Petroni, F.; Karpukhin, V.; Goyal, N.; Küttler, H.; Lewis, M.; Yih, W.-T.; Rocktäschel, T.; et al. Retrieval-Augmented Generation for Knowledge-Intensive NLP. In Advances in Neural Information Processing Systems 33 (NeurIPS 2020); Curran Associates: Red Hook, NY, USA, 2020; pp. 9459–9474. [Google Scholar]
- Konstantakis, M.; Michalakis, K.; Aliprantis, J.; Kalatha, E.; Moraitou, E.; Caridakis, G. A methodology for optimised cultural user personas experience-CURE architecture. In Proceedings of the 32nd International BCS Human Computer Interaction Conference, Belfast, UK, 4–6 July 2018. [Google Scholar]
- Calvo, R.A.; D’Mello, S. Affect Detection: An Interdisciplinary Review of Models, Methods, and Their Applications. IEEE Trans. Affect. Comput. 2010, 1, 18–37. [Google Scholar] [CrossRef]
- Brusilovsky, P.; Millán, E. User Models for Adaptive Hypermedia and Adaptive Educational Systems. Int. J. Artif. Intell. Educ. 2007, 17, 89–120. [Google Scholar]
- Falk, J.H.; Dierking, L.D. The Museum Experience Revisited; Routledge: London, UK, 2016. [Google Scholar]



| Criterion | Zarantonello [22] | Arkae–Vision [23] | Scene-Based XR [8] | Modular XR (JCH) [9] | ARIA |
|---|---|---|---|---|---|
| Framework type | Experience-oriented conceptual framework | System-oriented technical framework | Scene-based architectural framework | Modular system architecture | Conceptual and architectural framework |
| Primary purpose | Experience design and analysis | Development of interactive XR systems | Authoring and deployment of XR scenes | System structuring and modularization | End-to-end adaptive AR interpretation |
| Target CH context | General XR services | Museums and CH sites | Museum exhibits | Museums and heritage environments | GLAM institutions (in-situ AR) |
| Key components | Experience dimensions and value constructs | XR interaction modules and pipelines | Scene authoring and rendering components | Modular system layers | Sensing, user modeling, AI reasoning, DAM, AR delivery |
| Adaptivity model | Not specified | Interaction-driven (rule-based) | Not specified (static scenes) | Limited (predefined logic) | Real-time context- and affect-informed adaptation |
| AI integration | Not specified | Not specified | Not specified | Not specified | Yes (user modeling, RAG-based personalization) |
| Content grounding | Conceptual discussion only | Pre-authored content | Scene-level authoring | Repository-based content | Curator-approved knowledge sources (RAG) |
| Governance and ethics | Discussed at a conceptual level | Not specified/out of scope | Not specified/out of scope | Not specified/out of scope | Operationalized (GDPR, FATE, XAI) |
| Framework Component | Design Goal | Operational Mechanism | Indicative Validation Metrics |
|---|---|---|---|
| Affective Sensing Module | Engagement-aware adaptation | Multimodal proxy fusion (behavioral + contextual) | Latency, signal availability, adaptation trigger rate |
| Adaptive AI Engine (RAG) | Trustworthy personalization | Curated retrieval + controlled generation | Retrieval precision, narrative coherence, curator approval rate |
| Immersive AR Module | Usability and immersion | Context-driven content modulation | AR stability, interaction errors, dwell time |
| DAM Integration | Content integrity | FAIR-aligned metadata + semantic links | Metadata completeness, reuse rate |
| Ethical Governance Layer | Transparency and trust | Consent management + XAI overlays | Consent opt-in rate, explainability comprehension |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2026 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license.
Share and Cite
Konstantakis, M.; Iakovaki, E. ARIA: An AI-Supported Adaptive Augmented Reality Framework for Cultural Heritage. Information 2026, 17, 90. https://doi.org/10.3390/info17010090
Konstantakis M, Iakovaki E. ARIA: An AI-Supported Adaptive Augmented Reality Framework for Cultural Heritage. Information. 2026; 17(1):90. https://doi.org/10.3390/info17010090
Chicago/Turabian StyleKonstantakis, Markos, and Eleftheria Iakovaki. 2026. "ARIA: An AI-Supported Adaptive Augmented Reality Framework for Cultural Heritage" Information 17, no. 1: 90. https://doi.org/10.3390/info17010090
APA StyleKonstantakis, M., & Iakovaki, E. (2026). ARIA: An AI-Supported Adaptive Augmented Reality Framework for Cultural Heritage. Information, 17(1), 90. https://doi.org/10.3390/info17010090

