Next Article in Journal
Unveiling the Scientific Knowledge Evolution: Carbon Capture (2007–2025)
Previous Article in Journal
Enhancing the Effectiveness of Juvenile Protection: Deep Learning-Based Facial Age Estimation via JPSD Dataset Construction and YOLO-ResNet50
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Zero-Shot to Head-Shot: Hyperpersonalization in the Age of Generative AI

1
Centre for Data Analytics and Cognition, La Trobe University, Melbourne, VIC 3086, Australia
2
Core dna, Prahran, VIC 3181, Australia
3
Department of Computer Science, Virginia Commonwealth University, Richmond, VA 23284, USA
*
Author to whom correspondence should be addressed.
Appl. Syst. Innov. 2025, 8(6), 186; https://doi.org/10.3390/asi8060186 (registering DOI)
Submission received: 3 November 2025 / Revised: 24 November 2025 / Accepted: 27 November 2025 / Published: 30 November 2025

Abstract

Generative Artificial Intelligence (GenAI) is rapidly transforming industries and organizations through automation and augmentation. Personalization of human–system interaction is a key area that can be significantly advanced through the effective implementation of GenAI. GenAI, positioned as an intermediary between humans and systems, can transform the human experience from the pre-defined, conventional notions of personalization into a dynamic and integrated hyperpersonalization capability. This article presents the zero-shot-to-head-shot hyperpersonalization (Z2H2) framework, which aims to achieve this through the effective adoption of GenAI techniques. It is a domain-neutral framework of three incremental stages named zero-shot, few-shot, and head-shot that gradually increase the level of hyperpersonalization of the human–system interaction. The framework is further represented in a layered system design and the Z2H2 Data Modality Matrix (ZDMM), which systematically maps data types, AI capabilities, and personalization objectives for each stage. The capabilities of the framework are demonstrated in an educational setting, followed by an empirical evaluation using the Open University Learning Analytics Dataset (OULAD). Although this dataset only contains demographic and aggregated clickstream data, which is a subset of attributes relevant to the entire framework, the gradual development of zero-shot-to-head-shot hyperpersonalization is effectively demonstrated and validated on these student interactions.

1. Introduction

Recent innovations in GenAI are driving the rapid transformation of human–system interaction (HSI) from static and pre-defined personalization into dynamic and integrated hyperpersonalization that decreases the system complexity of each human interaction. This evolution has reshaped how we interact with systems by making every touchpoint more relevant and insightful [1]. Given its broad applicability across multiple domains, numerous definitions for personalization have emerged, each varying depending on the focus, implementation methods, intended outcome, and domain contexts. A generalized early definition of personalization [2] is defined as “whenever something is modified in its configuration or behavior by information about the user, this is personalization”. While personalization can occur in both online and offline contexts, it is generally referred to as custom virtual/online experiences using user data to meet individual needs [3]. Early implementations of personalization were primarily developed using rule-based systems, collaborative filtering, content-based recommendations, and user segmentation methods [4,5,6,7,8,9]. Recommender systems are foundational for delivering personalization, while collaborative filtering and content-based filtering methodologies are used to predict user preferences and selections. However, it is also observed that these systems typically prioritize system objectives at the expense of a personalized human experience [10]. These approaches consider the user as a member of a broader segment rather than as a unique individual, also referred to as “segments of one”. Although these techniques have been effective in conventional personalization, more recent HSIs expect highly relevant and individualized interactions that dynamically adapt to preferences and intentions. Thereby, conventional personalization systems face challenges in representing dynamic user behavior, real-time adaptability, and scalability and also tend to perform poorly in situations with limited user data, which is commonly referred to as the “cold start” problem [11]. Collectively, these limitations can be addressed through the transition from personalization to hyperpersonalization. Hyperpersonalization aims to provide a highly individualized human–system experience by adapting in real time to changing user behaviors, preferences, contexts, and intents. In this paper, we define hyperpersonalization as “the continuous adaptation of human–system interactions that significantly decrease the complexity of information, decision, and action to provide a continuously individualized experience of system use and operation”.
AI-driven personalization has been reported in the recent literature, where advancements in GenAI have accelerated its progression [12] towards hyperpersonalization. The extensive pretrained knowledge of LLMs offers the ability to generate highly personalized outputs even in cold start scenarios with minimal user input. This capability is enabled through various techniques, such as prompt engineering [13], in-context learning [14], retrieval-augmented generation (RAG) [15], and agentic AI [16]. These techniques allow the personalization systems to incorporate their pretrained knowledge, retrieved external information, contextual cues, and reasoning and planning capabilities to generate adaptive and relevant experiences. Although these advanced capabilities have significantly improved the personalization capabilities of modern systems, true hyperpersonalization remains unrealized. The primary gap in this regard is a persistent and evolving understanding of the user, which is the capacity of systems to remember, reason, and adapt continuously across sessions, contexts, and roles. Present methods are mainly session-bound and event-driven. Their level of personalization is based on immediate context rather than maintaining a lifelong user representation that captures cognitive and behavioral continuity. Furthermore, most approaches focus on narrow, task-specific applications rather than addressing hyperpersonalization as a framework that adapts in real time based on HSI depth and cognitive intent.
In this paper, we address this gap through the design and development of the Z2H2 framework, which consists of a three-layer architecture operating across three incremental stages. This paper makes the following three contributions: (1) the Z2H2 framework for hyperpersonalization, consisting of three incremental stages named zero-shot, few-shot, and head-shot, (2) the Z2H2 Data Modality Matrix (ZDMM), which formulates the four modalities of data, explicit, implicit, passive, and derived, for each of the three incremental stages, and (3) experimental validation of the framework’s capabilities in an educational setting for progressive HSI across the Z2H2 stages. The rest of this paper is organized as follows. Section 2 presents related work on personalization and hyperpersonalization using AI techniques, followed by Section 3, which outlines the proposed Z2H2 framework, detailing its layered architecture and stage-by-stage progression. Section 4 demonstrates the framework in a real-world educational setting, followed by an empirical evaluation in Section 5 using the OULAD dataset. Section 6 discusses limitations and ethical considerations and Section 7 concludes this paper.

2. Related Work

Personalization is a broad and widely studied area that spans numerous domains. In this section, we focus on the evolution of personalization towards hyperpersonalization and explore the existing literature, highlighting the role of AI, especially GenAI, in modern-day hyperpersonalization. Early-stage personalization systems were mainly rule-based or expert-driven. They used if–then rules or domain ontologies to personalize the user experience. The pioneering work of Goldberg et al. [17] introduced the idea of using collaborative filtering to make predictions about a user’s interest by looking into preferences from many other users, laying the foundation for recommender systems and personalization. In [18], authors introduce one of the early collaborative filtering systems, named GroupLens, which recommends news articles to users based on articles liked by similar users. However, collaborative filtering had limitations in addressing the cold start problem, and content-based filtering emerged as a solution. It focuses the individual user’s preferences and the content characteristics of items for recommendations. Hybrid methods such as Fab [19] combined collaborative and content-based features, and clustering-based approaches [20,21] were also widely used within this first generation of personalization systems. However, they were mainly bound to specific domains and showed limitations in adaptivity, where the updates were infrequent and not real-time.
With increasing data availability, incorporation of deep learning methodologies allowed personalization systems to capture nonlinear patterns and complex feature interactions. Models such as collaborative deep learning (CDL) [22] integrated deep autoencoders with matrix factorization to jointly learn item content and user preferences. Moreover, Deep Learning Recommendation Model (DLRM) [23] combined embedding-based categorical features with dense multilayer perceptrons or large-scale prediction tasks such as ad ranking. Recurrent Neural Networks (RNNs) were also widely explored in the domain [24,25] to capture session-based user content evolution over time. This allowed personalization systems to consider the order and the context of the prior events. Although early deep learning methods collectively contributed to the personalization domain, the majority of them mainly focused on optimizing recommendations. These systems often overlooked implicit cognitive signals, temporal intent evolution, and multimodal and cross-platform interactions, which are crucial elements for hyperpersonalization. Subsequent advancements in the domain have introduced a new set of approaches that appear more aligned with the objectives of hyperpersonalization. These include session-based recommenders, sequential user modeling architectures, multi-agent recommender systems, and large language model-driven personalization methods. These approaches extend beyond classical collaborative filtering and neural models by incorporating context relevancy, agentic collaboration, and generative modeling.
Session-based recommender systems such as GRU4Rec treat each session independently, aiming to predict the user’s next action based only on the current session’s behavior without relying on any prior user profile [26]. Subsequent improvements introduced attention mechanisms to capture the user’s intent within the session more effectively. For example, Neural Attentive Recommendation Machine (NARM) uses an attention encoder to emphasize a user’s main purpose in the session [27]. These session-based methods excel at adapting to a user’s short-term interests instantly, making them valuable when preferences shift quickly within a session. However, these approaches lack persistent user understanding, which causes them to ignore long-term preferences or past behavior patterns. The adaptation stops when the session ends, limiting any learning from one session to the next. Therefore, their short-term, session-only adaptation and absence of evolving user understanding distinguish them sharply from Z2H2’s persistent, multi-stage hyperpersonalization framework.
Sequential recommender systems extend personalization beyond session-based recommender systems by modeling the sequence of user interactions over time. This enables the systems to capture behavioral signals across multiple sessions rather than restricting personalization to a single interaction window. Transformer-based [28] architectures such as SASRec introduced self-attention mechanisms to dynamically weight past interactions, effectively modeling both short-range and long-range behavioral patterns [29]. Building upon this, BERT4Rec used a bidirectional transformer to model sequential data, which allowed these models to encode contextual information from both preceding and succeeding interactions [30]. These advanced integrations supported models to incorporate capabilities to recognize temporal trends, preference drift, and long-term behavioral structure. Although these systems maintain a form of temporal continuity by encoding multi-session behavior, their personalization mainly focuses on optimizing next-item prediction accuracy. This limits their ability to support richer user representation learning. Sequential recommender systems learn representations that reflect statistical co-occurrence rather than deeper user intent, goals, or identity, and unlike Z2H2, they also do not support progressive personalization stages and apply the same predictive pipeline regardless of whether the user is in a zero-shot, few-shot, or mature head-shot state.
LLM-driven personalization methods highlight the most recent shift within this trajectory. These systems incorporate the generative intelligence of LLMs to either directly generate recommendations or to improve parts of the recommendation pipeline. Chen et al. [31] argue that LLMs can be utilized as a general-purpose interface for personalization tasks. The abilities, like in-context learning and reasoning, in LLMs help interpret user requests, plan actions, and call external tools or APIs to fulfill user needs. Lyu et al. [32] introduce LLM-Rec, which is family of prompt templates that enrich the item description with paraphrases before feeding it to the recommender systems. Their work highlights that these enriched prompts benefit from in-context learning and outperform strong content-based baselines on the MovieLens-1M and Recipe datasets, without any additional user fine tuning. Extending prompt-level personalization, Tan et al. [33] propose PER-PCS, which is a modular parameter-efficient framework where users share small LoRA layers that can be recombined on demand. Furthermore, recent research has also explored improving classic models by injecting LLM-generated knowledge [34].
LLM-based personalization brings clear advancements towards realizing hyperpersonalization. Their broader pretrained knowledge helps personalization systems to handle zero-shot and few-shot scenarios, which then aligns with the Z2H2 positioning of zero-shot and few-shot hyperpersonalization where the system can start personalizing based on limited HSI data. Furthermore, LLMs introduce a form of cognitive reasoning into personalization where systems are able to infer user intents from context and conduct guided planning tasks. However, without special design, an LLM still struggles to remember a user across sessions due to the absence of evolving user representation learning. Each interaction is stateless unless we explicitly provide conversation history or employ external memory tools through RAG architectures [35,36]. This special design requirement motivated the emergence of multi-agent personalization systems. In the first instance within the hyperpersonaization trajectory, systems moved beyond recommendation, ranking, and content suggestion to interpret user intent, plan multi-step tasks, coordinate specialized agents, and execute actions on behalf of the user. Approaches such as MACRec [37] and Rec4AgentVerse [38] demonstrate how these multi-agent frameworks overcome limitations with single-model LLM personalization systems by introducing distributed intelligence, delegation, coordination, cross-agent verification, self-reflection, autonomous tool use, and task execution.
Modern-day multi-agent personalization systems stand as the closest class of technologies to true hyperpersonaization. However, their intelligence remains fragmented, as each agent focuses on only a partial and task-specific view of the user rather than contributing towards an evolving user understanding. Their coordination mechanisms are mainly task-driven and lack a progression pathway to gradually build up the required personalization depth across stages. Similar limitations have been highlighted in other real-time, multimodal cognitive systems. For example, authors demonstrate how situation awareness architectures must fuse multimodal data to extract controlling intent and forecast future system states [39]. This highlights the need for continuous, intent-aware, closed-loop modeling rather than isolated task execution. In contrast to the task-driven approach within multi-agent personalization systems, a hyperpersonalization system aims to be objective-driven, grounded in a lifelong understanding of the user’s evolving representation. Such systems focus on achieving continuity and adaptivity, progressively refining user models by integrating multimodal signals, contextual reasoning, and behavioral evolution. Z2H2 builds on the strengths of generative and agentic intelligence while directly addressing these limitations. Rather than reacting to isolated tasks, Z2H2 formalizes personalization as an ongoing, learning-driven process that strengthens in precision as user interactions accumulate. In the next section, we introduce the Z2H2 framework, which formalizes the evolution of personalization across multiple phases by incorporating GenAI technologies.

3. Methodology: The Proposed Z2H2 Framework

The proposed Z2H2 framework incrementally develops a user profile, beginning with zero-shot assumptions, advancing through few-shot behavioral adaptations, and finally evolving into head-shot cognitive precision. The zero-shot stage refers to the initial interaction stage, where the system relies heavily on limited prior data about a user to deliver personalization without observing any user behavior. The few-shot stage begins once the user starts interacting, which allows the system to adapt personalization using behavioral signals. The head-shot stage is reached when the system has accumulated sufficient data to infer the user’s objective, intent, and decision strategies by enabling deep personalization. The framework employs GenAI technologies across different stages of the hyperpersonalization session to deliver a deeply personalized experience unique to each user. It consists of three core components named the ZDMM, the cognitive layer, and the Hyperpersonalization Layer, which are responsible for data collection, profile creation, and personalized experience generation, respectively. GenAI systems have evolved from pretrained language models into autonomous multi-agent architectures. In our work, we aligned this generative intelligence within the Z2H2 framework, where each stage of the Z2H2 framework is associated with the following GenAI capabilities: reasoning, contextual understanding, and user intent alignment. Specifically, we focused on employing prompt engineering techniques for zero-shot, in-context learning (ICL) for few-shot, and agentic AI and RAG methodologies for head-shot hyperpersonalization. Figure 1 illustrates the three-layered Z2H2 framework.

3.1. The Z2H2 Data Modality Matrix

The data modality matrix is a foundational layer for the entire Z2H2 framework. Given the dynamic and objective-driven nature of hyperpersonalization, this layer focuses on the systematic integration of diverse data sources. This consists of all potential data points across a user journey, including evolving data streams, shifting objectives, and behavioral changes. Therefore, this evolving data foundation not only supports hyperpersonalization but also drives the framework’s ability to interpret intent and dynamically align its outputs with user goals. The matrix is composed of four data modalities: explicit, implicit, passive, and derived. The description and focus area of each modality are presented in Table 1.
To further demonstrate how these data modalities are incrementally accumulated and utilized within a hyperpersonalization session, we have aligned them across the zero-shot, few-shot, and head-shot stages. At the zero-shot stage, the focus is on the limited prior data available at the beginning of the session, which mostly consists of information that requires no or minimal user interaction. These early signals help the system deliver an initial personalization based on assumptions and generalizations. In the few-shot stage, the focus shifts to behavioral data, which helps the system to identify evolving preferences and adapt responses accordingly. At this stage, the system uses these signals to refine its understanding of the user to personalize interactions through behavioral alignment. In the head-shot stage, the system consists of sufficient cognitive-level data that reveals the user’s intent, urgency, and objectives. In Figure 2, we visualize the ZDMM to provide a unified view of the incremental data accumulation and utilization across modalities. The different types of data contributing to each modality and stage are discussed in detail in the following subsections.

3.2. Zero-Shot Stage

In the zero-shot stage of the Z2H2 framework, the ZDMM considers explicit inputs such as sign-up metadata, language preferences, and system-level inputs, as well as passive signals such as device type, browser metadata, location, time, and referrer URLs. Additionally, derived data types such as prebuilt personas, embedding matches using models like CLIP, BERT, or GPT, and knowledge graph-based priors are used to further enhance this early understanding. This diverse set of data collectively forms the groundwork for zero-shot hyperpersonalization in the absence of any user-specific behavioral data. Using the collected data, the system constructs an IP, which is represented as a dense vector p I . This serves as the initial representation of the user and is curated by fusing zero-shot data through embedding-based methods.
p I = f embed D explicit , D passive , D derived
where f embed ( · ) denotes a general embedding function that may be implemented using any suitable encoder architecture. These include simpler metadata-driven encoders or even transformer-based models such as BERT, CLIP, or GPT. Explicit and passive data points are converted into natural language descriptors (e.g., “mobile device on Saturday evening,” and “prefers English content”) and passed through transformer-based encoders to produce embeddings. Derived data, such as persona embeddings, are retrieved by matching the user’s explicit and passive inputs against a bank of prebuilt persona profiles using cosine similarity over existing persona vectors. To further enhance the p I , neighborhood aggregation or graph attention methods are used to identify semantically related nodes (regional preferences and related topics) within knowledge graphs, which are then encoded into graph-aware embeddings. Furthermore, social trends are encoded by embedding textual representations of aggregate behaviors (e.g., “most learners in this region prefer short weekend courses”). All these embeddings are then concatenated using embedding fusion techniques to form the final p I . Representing IP as an embedding profile ensures memory efficiency, supports vector-based operations and fusion techniques, and also enables effective integration with contemporary GenAI systems.
Once p I is constructed, the GenAI-driven hyperpersonalization is initiated by conditioning an LLM on this initial profile using prompt engineering techniques, as shown in (2). Here, y is the personalized output, and it can take various forms such as natural language responses, structured content, or even interface configurations. C sys acts as the system prompt, and it sets the behavior of the model by defining the tone, role, and other constraints. u 0 represents any available user inputs. p I is included in the prompt by transforming the embedding vector into a set of semantic descriptions or persona attributes that the LLM can interpret. Thereafter, these descriptors are concatenated with C sys and u 0 to form the complete prompt. Also, p I can be directly integrated with the LLM’s token embedding space by using soft prompts or adapter-based conditioning. This allows the LLM to align the model’s generation with the semantic attributes encoded in p I [40].
y = LLM prompt ( C sys , u 0 , p I )
Advanced prompting strategies such as persona-based prompting, Chain-of-Thought (CoT) prompting, and contextual prompting can be used at this stage to refine the personalization quality [41]. Through this zero-shot hyperpersonalization, even first-time users can experience a level of personalization, as the system is capable of using the limited prior data to generate recommendations, personalized content, and dynamic UI layouts, thus effectively addressing the cold start problem.

3.3. Few-Shot Stage

This stage of the Z2H2 framework begins once the user initiates active interactions with the system. At this phase, the ZDMM considers explicit data such as user-provided inputs, including search queries, applied filters, bookmarking, or other preferences. Implicit data mainly emerges from user behavior patterns and consists of clicks, scroll depth, dwell time, and navigation patterns, which indirectly reveal the user’s short-term preferences and intent. Passive data, such as location changes, device state transitions, or fluctuating network conditions, will decide the type of content or level of personalization delivered at this phase. Finally, derived data becomes more relevant as interaction data helps summarize high-level semantic patterns, such as inferred intent behind search queries and other behavior patterns. The integration of these four modalities ensures the transition of the Z2H2 framework from zero-shot assumptions to behavior-driven few-shot hyperpersonalization.
When compared to the zero-shot phase, which is based mainly on limited prior data, the few-shot cognitive layer incrementally constructs a behavioral profile (BP), which can be denoted as the vector form p B ( t ) . It reflects the user’s evolving preferences, interests, and decision-making patterns during the few-shot phase. This helps the system to deliver real-time adaptation of personalization using ICL and embedding fusion techniques. Each user interaction is transformed into a behavioral embedding ( e t ) using a transformer-based encoder, as denoted in (3), where u t is the i th user interaction in natural language form. Then, these interaction embeddings are accumulated and fused with the previous BP state using an exponential moving average, as in (4). Here, 0 < α < 1 controls the balance between past and the most recent user interactions. To maintain the impact from the context of the zero-shot phase, the BP is combined with the ( p I ) using a variable γ , where 0 < γ < 1 , and it balances prior assumptions with new behavioral insights, as depicted in (5). p B * ( t ) denotes the enriched behavior profile, which combines p I with the evolving behavior profile p B ( t ) .
e t = f embed ( u t )
p B ( t + 1 ) = α p B ( t ) + ( 1 α ) e t + 1
p B * ( t ) = γ p I + ( 1 γ ) p B ( t )
In the few-shot phase, the GenAI layer utilizes this enriched vector p B * ( t ) as the input to the LLM by embedding it directly or serializing it into semantic descriptors. The prompt is extended by using both the user’s recent interaction history u 0 : t and the behavioral context vector p B * ( t ) to achieve ICL, as shown in (6). Here, u 0 : t = { ( u 0 , r 0 ) , ( u 1 , r 1 ) , , ( u t , r t ) } captures the sequence of user interactions u i and corresponding system replies r i . In this setting, ICL guides the LLM by embedding the interaction and response pairs as few-shot examples within the prompt. The final prompt can be expressed as shown in (7), where ⊕ denotes concatenation and Format ( u i , r i ) represents the structured encoding of each user interaction and the corresponding response. Serialize ( p B * ( t ) ) transforms the high-dimensional embedding p B * ( t ) into a natural language summary or structured descriptors so that the LLM can interpret and condition on the profile semantics.
y = LLM ( P ) = LLM prompt ( C sys , u 0 : t , p B * ( t ) )
P = C sys i = 0 t Format ( u i , r i ) Serialize ( p B * ( t ) )
By including both behavioral examples and serialized profile information in the final LLM prompt, the LLM is capable of inferring evolving user preferences and dynamically adapting its output, such as recommendations, personalized actions, summaries, or even system layouts. This allows the system to deliver adaptive personalization to the end user by continuously refining outputs as the session evolves.

3.4. Head-Shot Stage

The transition into the head-shot stage is based on the accumulation of sufficient data to infer the user’s cognitive intent and long-term objectives. This phase focuses on delivering deep and precise personalization. At this stage, the system already consists of zero-shot and few-shot data, and it collectively uses this data along with head-shot phase data to construct the Cognitive Intent Profile (CIP), represented as p C ( t ) . Explicit data such as feedback, survey responses, goal declarations, or chatbot interactions become available, providing direct insights into the user’s long-term objectives. Implicit data, such as sequence modeling of multi-step user interactions, recurrence patterns, and decision-making behaviors observed over time, indicate user intent. Passive data evolves to include multimodal inputs and tool usage metadata, which help the system understand the user’s environment. Finally, derived data supports the system by providing inferred user urgency levels, decision strategies, and outcomes from agentic AI reasoning to build a more precise and cognitively aligned CIP. The head-shot cognitive layer is responsible for constructing the CIP, represented as p C ( t ) in its vector form. This layer abstracts all head-shot data from the ZDMM and goes beyond behavioral embeddings by incorporating cognitive embeddings, as defined in (8). In this context, u t denotes the t t h interaction, D h e a d represents the head-shot data modalities, and f c o g is an embedding function that transforms user interactions and head-shot data into a cognitive embedding vector. To implement f c o g , the system utilizes transformer-based LLMs for textual data, multimodal encoders such as CLIP for non-textual data, and knowledge graph-based embeddings to capture the relationships between user objectives and domain concepts.
c t = f cog ( u t , D head )
The CIP evolves over time as new cognitive embeddings are observed, and p C ( t ) is updated incrementally, similarly to the BP in the few-shot phase. This update is defined in (9), where 0 < β < 1 controls the balance between historical cognitive signals and new head-shot insights. Furthermore, to ensure that all relevant data from the zero-shot and few-shot phases are incorporated into the head-shot phase, p C ( t ) is fused with the enriched BP, p B * ( t ) , to form the complete and enriched p C * ( t ) , as formally presented in (10). Here, 0 < λ < 1 controls the weighting between historical behavioral context and inferred cognitive intent. This two-step CIP construction ensures that it captures both historical user behavior patterns from p B * ( t ) and other related, real-time information required for accurate cognitive modeling.
p C ( t + 1 ) = β p C ( t ) + ( 1 β ) c t + 1
p C * ( t ) = λ p B * ( t ) + ( 1 λ ) p C ( t )
Once p C * ( t ) is finalized, this complete representation of the user is utilized by the system, along with GenAI capabilities, to generate precise, objective-driven hyperpersonalization for the user, as shown in (11). The system context, C sys , provides high-level task instructions to the LLM, while u 0 : t represents the user interaction history, ensuring that the LLM has a clear understanding of prior interaction steps. Additionally, p C ( t ) consists of serialized natural language descriptors. RAG further contributes real-time external content, which helps enhance the input prompt [42]. Agentic AI plays a crucial role in this phase by leveraging advanced agentic capabilities that enable hyperpersonalization systems to perform objective-driven decision making.
y = LLM prompt ( C sys , u 0 : t , p C * ( t ) , R )
Advanced reasoning and task decomposition capabilities help the system understand the “why” behind user actions, predict future needs, and identify decision-making strategies, similarly to intent-driven session-based reasoning approaches [43]. Agentic AI leverages the reasoning capabilities of LLMs to decompose complex personalization tasks into reasoning chains and sub-goals. Frameworks like AutoGPT, LangChain, and LangGraph demonstrate these capabilities through structured agent planning and decision making. Head-shot hyperpersonalization also benefits from multimodal agentic AI capabilities, which enable processing and generation across various modalities such as text, audio, image, and video. Multimodal generative models further allow the system to respond in the appropriate modality, producing images, audio, or videos personalized to the user. In the head-shot phase, agentic AI extends beyond response generation to goal-oriented actions. Tool calling capabilities enable the system to execute external services or APIs to fulfill complex user goals. Furthermore, as hyperpersonalization reaches greater cognitive depth, interactions become increasingly human-like. Socially aware AI agents (avatars or personas) can maintain continuity across conversation turns and adapt their tone and behavior based on the user’s CIP. Additionally, recent advancements such as Model Context Protocol (MPC) can also be incorporated in this layer to further enhance modular tool orchestration and external integrations. Together, these advanced GenAI capabilities will contribute towards a complete system that incrementally improves its understanding of the user to deliver precise, objective-driven, and user-centric hyperpersonalization.

4. Z2H2 in Education

Aligned with the progressive and gradual nature of learning, the Z2H2 framework is demonstrated for the education sector to personalize the student learning experience, as illustrated in Figure 3. In this example, a learner maintains the objective of finding a suitable course in the GenAI domain that matches their preferences. The referral may come from a search engine, an LLM suggestion, or a UTM-based marketing campaign. The data available at the zero-shot stage is defined by the ZDMM, and it includes explicit inputs such as language preferences or sign-up metadata if available. Passive data related to the device, location, time of access, etc., and derived data such as regional educational trends and prebuilt persona embeddings help the system to establish an initial understanding of the learner without any active input.
The cognitive layer processes these modalities through embedding techniques and similarity checks against prebuilt personas to construct the p I . For example, if the learner interaction happens on a Saturday night and the learner device is a mobile phone, the system may identify the persona of a “casual weekend learner”. This helps the GenAI layer to deliver personalization even at the zero-shot phase. The personalized outputs may include course recommendations, homepage layout suggestions, onboarding headlines, and more. More technically advanced personalization can be integrated by integrating multimodal capabilities such as personalized banner generation using text-to-image GenAI models (e.g., DALL·E and Stable Diffusion) and personalized chatbots. Once the user starts to interact with the platform, it transitions to the few-shot phase.
In this phase, the data layer captures meaningful behavior inputs such as search queries (GenAI courses), applied filters (short projects), and other engagement inputs such as course previews, scroll depth, or time spent exploring specific courses (opened Course A with 80% scroll depth and skipped Course B). In the cognitive layer, these behavioral interactions are semantically embedded and fused with the p I to refine the p B . Thus, the cognitive layer is capable of inferring a more meaningful persona like “beginner GenAI enthusiast with weekend project focus” to match the user’s few-shot behavioral inputs. In the GenAI layer, ICL is used to embed recent interaction and response pairs into the LLM to support context-aware recommendations without any additional training. For example, if a learner previously searched “GenAI course for beginners” ( u 1 ) and received recommendations [Course A, B, C] ( r 1 ), this helps the LLM to understand the learner’s evolving preferences and deliver more targeted course suggestions and personalized outputs. The personalized outputs at this phase will be more refined and aligned towards the user’s short-term intent. The current page layout may also be dynamically adapted to prioritize frequently interacted topics or sections that align with the BP.
Once sufficient behavioral data is collected to infer the learner’s long-term intent, the platform transitions into the head-shot phase. At this phase, the platform integrates a full representation of the user along with the historical interaction sequences to construct the CIP ( p C * ). The CIP considers the learner’s declared goals, final course selections, learning style, urgency level, and commitment data to infer a more accurate cognitive interpretation of the learner. External knowledge is also retrieved through RAG to enhance the context with real-time insights such as trending GenAI case studies. In the GenAI layer, agentic frameworks apply reasoning and tool calling capabilities to decompose learner objectives into subtasks, invoke relevant APIs for tasks such as career path generation and course enrollment, and personalized discount generation. Agentic frameworks such as LangChain, AutoGPT, and LangGraph can be employed in this phase. This final stage delivers a hyperpersonalized experience to the learner, which includes a personalized course interface, one-click enrollment with pre-filled preferences, and a post-enrollment roadmap that includes certification and career path recommendations. At the end of the session, both the user’s goal and the system’s goals have been achieved. The learner finds the most suitable GenAI course aligned with their preferences, and the system successfully enrolls a new learner to the platform, completing this progressive hyperpersonalization session.
Although demonstrated for an educational setting, the proposed Z2H2 framework is equally applicable to any other hyperpersonalization setting across diverse sectors. For example, in industrial settings of human–system interaction, such as factory automation or cyber–physical systems, the framework will be represented as zero-shot for novice operators or repetitive functions, few-shot for intermediate operators and tasks, and head-shot for expert users and operations. The three layers and corresponding components are sector-agnostic and generalizable to any other application setting that involves human–system interaction with increasing complexity.

5. Experiments and Results

Drawing on the demonstration in an educational setting, this section presents a further empirical evaluation using the Open University Learning Analytics Dataset (OULAD) [44]. This dataset only contains a subset of attributes and semantics relevant to the proposed framework; therefore, experiments were conducted within the well-defined settings of a recommendation task and evaluated using the Hit Rate at 5 (HR@5) metric. Remaining capabilities of the framework will be evaluated as future work in a comprehensive experimental setting.

5.1. Experimental Setting

The OULAD dataset contains data from courses presented at the Open University (OU) and includes demographic data together with aggregated clickstream data of students’ interactions in the Virtual Learning Environment (VLE). In this experiment we focused on one course presentation (FFF-2013J), which contains a large student population and a rich set of VLE interactions suitable for this evaluation task. Furthermore, to ensure there is enough data to build reliable user representations at each stage, we kept only the students who had at least 40 VLE interactions and at least two assessment submissions for the selected course presentation.
Using the resulting cohort of 1697 student, we evaluated the Z2H2 framework in three stages. For the zero-shot profile, we used demographic and registration time attributes, which were available before any VLE interaction occurred. The few-shot profile was constructed using the early VLE interaction sequence (first 10 interactions). The head-shot profile used the complete interaction history up to the submission of the second assignment along with key assessment attributes. Within this experiment, we framed personalization as a recommendation task and evaluated how well each stage could predict the student’s next learning resource as the system gained more information across the stages. This setup allowed us to assess how personalization improves when more evidence becomes available and how the user understanding deepens across the Z2H2 stages. At each stage, recommendations were generated by ranking all learning resources based on the cosine similarity between the learner’s profile embedding and the corresponding resource embeddings. Across all stages, recommendation quality was measured using the Hit Rate at 5 (HR@5), which is a widely used metric for evaluating the accuracy of Top-K recommendations [45]. HR@5 measures whether at least one ground truth item appears in the Top-5 recommendations. To contextualize performance, a random recommendation baseline was computed by sampling Top-5 items uniformly from the full resource pool of the module. Random recommendation baselines are often used within recommender systems research [46] as a simple point of comparison that reflects chance-level performance, providing a clear reference point for understanding how much information is gained at each Z2H2 stage.

5.2. Results

The zero-shot stage tests for limited, pre-behavioral information alone can help predict the initial VLE interactions of the student. For each student, an initial profile vector p I was constructed using only features available prior to engagement with the module. These included gender, region, highest education level, IMD band, age band, disability status, and the student’s relative registration time. These attributes were then encoded, normalized, and projected into a shared 16-dimensional latent space through a random Gaussian transformation. Zero-shot recommendations were generated by computing cosine similarity between p I and the embeddings of all resources available in the selected module presentation. We considered the ground truth for zero-shot as the first five VLE interactions by the student. The zero-shot evaluation achieved an HR@5 of 0.0896, significantly outperforming the random baseline of 0.0354. This indicates that purely static demographic and enrollment features carry meaningful information about a learner’s early behavioral tendencies within the VLE.
The few-shot stage evaluates the Z2H2 framework once behavioral data becomes available through early interactions. Unlike the zero-shot stage, which relies entirely on limited prior data encoded in the initial profile, the few-shot phase incorporates the first ten observed VLE interactions to construct the p B . The first n = 10 interactions formed the context window used for behavioral modeling, while the subsequent m = 5 interactions served as the ground truth for evaluation. Each interaction in the context window was mapped to its corresponding resource embedding e t using the 16-dimensional latent representation defined in the experimental setup. Thereafter, following Equation (4), these behavioral signals were accumulated into a smoothed behavioral representation using the exponential moving average update rule. After processing the first ten interactions, the p B ( n ) was fused with the initial profile p I according to Equation (5), producing a few-shot representation p B * that balances zero-shot priors with early behavioral information. Next, this fused representation was used to rank all available resources in the module via cosine similarity, producing Top-5 recommendations for each learner. Using α = 0.3 for the behavioral update and γ = 0.2 for fusing the BP with the initial profile, the few-shot stage achieved an HR@5 of 0.1402. When compared to the zero-shot stage (HR@5 = 0.0896), the few-shot configuration improves hit rate despite operating on a more challenging prediction target. In zero-shot, the model predicts a learner’s very first five interactions, which are normally dominated by common entry point resources. In contrast, the few-shot stage must predict interactions that occur deeper in the sequence, where behavior is more diverse and personalized. The strong improvement over zero-shot under these harder conditions demonstrates that even a small amount of observed behavior significantly enhances personalization beyond what can be achieved from static and limited prior information alone.
The head-shot phase integrates three distinct information sources: static priors from the initial profile, an extended history of VLE interactions, and head-shot signals inferred from early assessment attributes. The second assessment was selected as the head-shot boundary, as later assessments would greatly reduce the cohort size and leave too few future interactions to predict. All VLE interactions prior to this timestamp were considered for the behavioral context for head-shot representation and the next m h = 5 interactions formed the ground truth for evaluation. The behavioral component for the head-shot stage was reconstructed by considering and processing the full sequence of VLE interactions up to the head-shot boundary. We used the same behavioral smoothing parameter α = 0.3 in accordance with Equation (4) and it produced an extended behavioral representation p B , head . This was then fused with the initial profile using Equation (5) and γ = 0.2 to form the enriched behavioral vector p B , head * . We then added head-shot features taken from the first two assessments such as type, weight, score, pass or fail outcome, and submission timing by projecting these attributes into the shared embedding space and accumulating them using the smoothing rule in Equation (9) with α = 0.3 . Finally, the behavioral representation and the assessment-based representation were combined using Equation (10) with λ = 0.2 to form the complete head-shot profile. Recommendations were generated by ranking all resources using cosine similarity with this profile, and the head-shot stage achieved an HR@5 of 0.2024. This improvement is also noticeably higher than the few-shot HR@5 of 0.1402, showing that the extended behavior and the head-shot signals provides a more mature and anticipatory representation of learner intent. Collectively, these findings confirm that the Z2H2 framework deepens personalization in a staged, evidence-driven manner, with the head-shot stage delivering the most precise level of personalization, thus establishing its contribution as a key enabler in achieving hyperpersonalization.

5.3. Parameter Sensitivity Analysis

Sensitivity analysis was conducted to further evaluate the influence of behavioral hyperparameters on performance, using α { 0.3 , 0.5 , 0.7 , 0.9 } and γ { 0.2 , 0.5 , 0.8 } . Figure 4a shows how HR@5 changes as these parameters vary. According to Equations (4) and (5), the parameter α controls how fast the behavioral profile changes and the parameter γ controls how much the model trusts the initial static profile p I versus the learned behavioral profile p B ( t ) . The curves clearly illustrate that lower values of both parameters lead to better performance. A low α allows the profile to adapt quickly to the learner’s most recent actions, which is important in the early stages where only a short sequence of interactions is available. A low γ reduces dependence on the static zero-shot information and increases the contribution of actual behavior, which strengthens the recommendation accuracy. The γ = 0.8 curve shows a slight upward drift for larger α values, which occurs because the model relies heavily on the static profile and only weakly incorporates behavior signals. These patterns align with the intended design of the few-shot stage. The results indicate that once behavioral evidence becomes available, the model naturally reduces its dependence on limited and static zero-shot information and grounds personalization on the user’s behavioral information.
Similarly, to evaluate how the head-shot hyperparameters affect performance, we evaluated β { 0.3 , 0.5 , 0.7 , 0.9 } and λ { 0.2 , 0.5 , 0.8 } . Figure 4b shows the resulting HR@5 trends. In the head-shot scenario, β governs how strongly the head-shot features are smoothed over time, and λ controls the balance between the extended behavioral profile and the additional head-shot features. The curves show that HR@5 decreases as β increases, meaning that heavier smoothing weakens the influence of the head-shot component and reduces its contribution to the final profile. The fusion weight also has a clear effect. Lower λ values perform well only when β is small, after which accuracy drops as the head-shot features become over-smoothed. These patterns are different from the few-shot stage because head-shot combines long-range behavior signals and assessment-based head-shot signals. This also reflects a limitation of the OULAD dataset, where the available head-shot features are high-level and limited and therefore not strong enough to fully support the intended head-shot representation. Taken together, these patterns show that the head-shot stage remains grounded in long-range behavior, with the head-shot features providing a lighter, complementary adjustment rather than acting as the dominant signal within the OULAD experiment setting.

6. Limitations and Ethical Considerations

Effective hyperpersonalization requires processing each user action instantly, tracking evolving intent, and maintaining a dynamic session state. Meta’s “Scaling User Modeling” work shows that offline embeddings can become stale when user features or intent shift. Therefore, challenges such as scalable infrastructure, multimodal data fusion, session-based context tracking, memory management, and high computational costs must be addressed while developing such systems. While users expect a highly personalized experience within their digital journey, they are greatly concerned about sharing their user data with such digital platforms. This is a popular scenario in the personalization domain known as the personalization–privacy paradox. Hyperpersonalization relies on a large collection of user data and thus raises various privacy and regulatory concerns while being implemented. Regulations like GDPR and the Australian Privacy Act 1988, and federal-level privacy acts such as the California Consumer Privacy Act (CCPA), explicitly define how personal data should be stored and handled [47,48]. According to these acts, the systems must provide transparent consent flows that allow users to control and manage their data. Thus, secure data collection, management, encryption, access control, and maintaining visibility are crucial factors to be considered while implementing hyperpersonalization systems. These systems should adopt “privacy by design” principles to address this personalization–privacy paradox and to deliver a trustworthy and personalized experience to the end user. Hyperpersonalization must also consider social concerns. Filter bubbles and echo chambers are risks of excessive personalization. These prevent the user from exposure to different perspectives and often lead to polarized opinions and reduced openness.
Furthermore, the inclusion of generative intelligence also introduces its own set of challenges and limitations into the hyperpersonalization pipeline. Large-scale pretrained LLMs and LLM-driven agentic systems are prone to hallucinate factual details, propagate hidden biases from their training data, and generate recommendations based on opaque internal reasoning. Within a hyperpersonalization pipeline, these issues can amplify preexisting filter bubbles by over-fitting to inferred preferences or by reinforcing biased correlations present in historical data. Within the Z2H2 architecture, these risks can be mitigated at multiple layers. At the ZDMM layer, explicit, implicit, passive, and derived signals can be validated, cross-checked, and diversified to provide the model with precise structured signals, thereby limiting hallucinated intent. At the cognitive layer, the profile fusion rules provide a natural mechanism for fairness and debiasing. The hyperparameters α , β , γ , and λ allow the system to detect inconsistent or low-confidence signals at each stage and down-weight them, preventing inferred bias from dominating the evolving user profiles. The contribution of these parameters in the fusion mechanism is examined in detail in Section 5. At the GenAI layer, grounding mechanisms such as RAG, rule-based guardrails, transparent reasoning chains, and agentic self-verification can help reduce factual as well as faithfulness hallucinations. This layered and staged architecture helps Z2H2 to approach hyperpersonalization as a more balanced and open personalization trajectory, thereby reducing the likelihood of isolating users within filter bubbles and echo chambers. Collectively, these methodologies can be employed to mitigate LLM-based hallucinations as well as system-based limitations such as bias, filter bubbles, and echo chambers.

7. Conclusions

This paper presented the Z2H2 hyperpersonalization framework, consisting of three incremental stages named zero-shot, few-shot, and head-shot that gradually increase the level of hyperpersonalization of the HSI. The framework is further unpacked into a layered system design and the Z2H2 Data Modality Matrix (ZDMM), which systematically maps data types, AI capabilities, and personalization objectives for each stage. Compared to a one-size-fits-all approach, Z2H2 incorporates multiple levels of user engagement and data availability for delivering a scalable, hyperpersonalized experience to the end user. The Z2H2 framework design focuses on the technical, privacy, and ethical limitations that can be listed as some of the major challenges when developing hyperpersonalization systems. From a technical perspective, the layered and staged architecture gives the framework flexibility to act as independent modules. This modularity allows the system components to scale individually and be deployed incrementally. The Z2H2 framework supports multimodal data fusion through its structured data modality matrix, allowing integration of behavioral, contextual, and inferred data streams. On the privacy side, Z2H2 follows a privacy-by-design principle and ensures the user data is processed only with explicit consent, securely stored, and in full compliance with the data protection regulations described in the previous section. The framework promotes transparent consent flows, encrypted storage, and user-controlled data access to allow users to have control over their personal information. Moreover, Z2H2 addresses social concerns due to over-personalization, listed as filter bubbles, echo chambers, and loss of user agency, by incorporating design safeguards that promote content diversity, transparency in decision making, and override capabilities. The agentic AI components in the head-shot phase ensure that users can drive their own experience, which enables a more open, fair, and trustworthy hyperpersonalized experience.
The experimental validation of the Z2H2 framework stages demonstrated that each stage contributes to meaningful improvements. These evaluations confirm the staged progression and illustrate that user representation becomes more precise as more evidence becomes available and relevant across the zero-, few-, and head-shot stages. The experiments only focused on evaluating the representational core of the Z2H2 framework to validate the Z2H2 profile construction process prior to introducing the complexity of generative models for hyperpersonalization delivery. As future work, the generative and agentic components of the Z2H2 framework will be evaluated by integrating RAG, reasoning chains, tool execution, and multimodal agent workflows across different domains. A further research direction is conducting user-centric social studies such as controlled user evaluations to understand the user acceptance level of the personalization produced by the Z2H2 framework and how well the system aligns with real user expectations.

Author Contributions

Conceptualization, S.S., N.M., H.M., D.D.S. and M.M.; Methodology, K.D., S.S., S.J., N.M. and D.D.S.; Software, K.D. and N.M.; Validation, K.D., N.M. and M.M.; Formal analysis, K.D., S.J. and H.M.; Investigation, S.S. and S.J.; Resources, S.J. and M.M.; Data curation, H.M.; Writing—original draft, K.D., S.S., S.J., N.M., H.M., D.D.S. and M.M.; Visualization, H.M.; Supervision, S.S., D.D.S. and M.M.; Project administration, D.D.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The original data presented in the study are openly available in https://analyse.kmi.open.ac.uk/open-dataset (accessed on 26 November 2025).

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Morton, F.; Benavides, T.T.; González-Treviño, E. Taking Customer-Centricity to New Heights: Exploring the Intersection of AI, Hyper-Personalization, and Customer-Centricity in Organizations. In Smart Engineering Management; Machado, C., Davim, J.P., Eds.; Springer International Publishing: Cham, Switzerland, 2024; pp. 23–41. [Google Scholar] [CrossRef]
  2. Searby, S. Personalisation—An Overview of its Use and Potential. BT Technol. J. 2003, 21, 13–19. [Google Scholar] [CrossRef]
  3. Chandra, S.; Verma, S.; Lim, W.M.; Kumar, S.; Donthu, N. Personalization in Personalized Marketing: Trends and Ways Forward. Psychol. Mark. 2022, 39, 1529–1562. [Google Scholar] [CrossRef]
  4. Chen, X.; Deng, H. Research on Personalized Recommendation Methods for Online Video Learning Resources. Appl. Sci. 2021, 11, 804. [Google Scholar] [CrossRef]
  5. Idrissi, L.E.Y.E.; Akharraz, I.; Ahaitouf, A. Personalized E-Learning Recommender System Based on Autoencoders. Appl. Syst. Innov. 2023, 6, 102. [Google Scholar] [CrossRef]
  6. Choi, O.; Han, S.Y. Personalization of Rule-Based Web Services. Sensors 2008, 8, 2424–2435. [Google Scholar] [CrossRef]
  7. Lu, J.; Wu, D.; Mao, M.; Wang, W.; Zhang, G. Recommender system application developments: A survey. Decis. Support Syst. 2015, 74, 12–32. [Google Scholar] [CrossRef]
  8. Liu, D.R.; Lai, C.H.; Lee, W.J. A hybrid of sequential rules and collaborative filtering for product recommendation. Inf. Sci. 2009, 179, 3505–3519. [Google Scholar] [CrossRef]
  9. Adomavicius, G.; Huang, Z.; Tuzhilin, A. Personalization and Recommender Systems. In INFORMS Tutorials in Operations Research; INFORMS: Catonsville, MD, USA, 2014; pp. 55–107. [Google Scholar] [CrossRef]
  10. Kazienko, P.; Cambria, E. Toward Responsible Recommender Systems. IEEE Intell. Syst. 2024, 39, 5–12. [Google Scholar] [CrossRef]
  11. Lika, B.; Kolomvatsos, K.; Hadjiefthymiades, S. Facing the cold start problem in recommender systems. Expert Syst. Appl. 2014, 41, 2065–2073. [Google Scholar] [CrossRef]
  12. Walch, K. How Generative AI is Driving Hyperpersonalization. Forbes. 15 July 2024. Available online: https://www.forbes.com/sites/kathleenwalch/2024/07/15/how-generative-ai-is-driving-hyperpersonalization/ (accessed on 1 July 2025).
  13. White, J.; Fu, Q.; Hays, S.; Sandborn, M.; Olea, C.; Gilbert, H.; Elnashar, A.; Spencer-Smith, J.; Schmidt, D.C. A Prompt Pattern Catalog to Enhance Prompt Engineering with ChatGPT. arXiv 2023, arXiv:2302.11382. [Google Scholar] [CrossRef]
  14. Dong, Q.; Li, L.; Dai, D.; Zheng, C.; Ma, J.; Li, R.; Xia, H.; Xu, J.; Wu, Z.; Chang, B.; et al. A Survey on In-Context Learning. In Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing (EMNLP 2024), Miami, FL, USA, 12–16 November 2024; pp. 1107–1128. [Google Scholar] [CrossRef]
  15. Siriwardhana, S.; Weerasekera, R.; Wen, E.; Kaluarachchi, T.; Rana, R.; Nanayakkara, S. Improving the Domain Adaptation of Retrieval Augmented Generation (RAG) Models for Open Domain Question Answering. Trans. Assoc. Comput. Linguist. 2023, 11, 1–17. [Google Scholar] [CrossRef]
  16. Acharya, D.B.; Kuppan, K.; Divya, B. Agentic AI: Autonomous Intelligence for Complex Goals—A Comprehensive Survey. IEEE Access 2025, 13, 18912–18936. [Google Scholar] [CrossRef]
  17. Goldberg, D.; Nichols, D.; Oki, B.M.; Terry, D. Using Collaborative Filtering to Weave an Information Tapestry. Commun. ACM 1992, 35, 61–70. [Google Scholar] [CrossRef]
  18. Resnick, P.; Iacovou, N.; Suchak, M.; Bergstrom, P.; Riedl, J. GroupLens: An Open Architecture for Collaborative Filtering of Netnews; Working Paper ccsw p165; MIT Center for Coordination Science: Cambridge, MA, USA, 1994. [Google Scholar]
  19. Balabanović, M.; Shoham, Y. Fab: Content-Based, Collaborative Recommendation. Commun. ACM 1997, 40, 66–72. [Google Scholar] [CrossRef]
  20. Ungar, L.H.; Foster, D.P. Clustering Methods for Collaborative Filtering. In Proceedings of the AAAI Workshop on Recommendation Systems, Madison, WI, USA, 26–30 July 1998; pp. 114–129. [Google Scholar]
  21. Schafer, J.B.; Konstan, J.; Riedl, J. Recommender Systems in E-Commerce. In Proceedings of the 1st ACM Conference on Electronic Commerce (EC ’99), Denver, CO, USA, 3–5 November 1999; pp. 158–166. [Google Scholar] [CrossRef]
  22. Wang, H.; Wang, N.; Yeung, D.Y. Collaborative Deep Learning for Recommender Systems. In Proceedings of the 21st ACM SIGKDD International Conference on Knowledge Discovery and Data Mining (KDD 2015), Sydney, NSW, Australia, 10–13 August 2015; pp. 1235–1244. [Google Scholar] [CrossRef]
  23. Naumov, M.; Mudigere, D.; Shi, H.J.M.; Huang, J.; Sundaraman, N.; Park, J.; Wang, X.; Gupta, U.; Wu, C.J.; Azzolini, A.G.; et al. Deep Learning Recommendation Model for Personalization and Recommendation Systems. arXiv 2019, arXiv:1906.00091. [Google Scholar] [CrossRef]
  24. Kelen, D.; Daróczy, B.; Ayala-Gómez, F.; Ország, A.; Benczúr, A. Session Recommendation via Recurrent Neural Networks over Fisher Embedding Vectors. Sensors 2019, 19, 3498. [Google Scholar] [CrossRef]
  25. Hidasi, B.; Karatzoglou, A.; Baltrunas, L.; Tikk, D. Session-Based Recommendations with Recurrent Neural Networks. arXiv 2016, arXiv:1511.06939. [Google Scholar] [CrossRef]
  26. Hidasi, B.; Karatzoglou, A. Recurrent neural networks with top-k gains for session-based recommendations. In Proceedings of the 27th ACM International Conference on Information and Knowledge Management, Torino, Italy, 22–26 October 2018; pp. 843–852. [Google Scholar]
  27. Li, J.; Ren, P.; Chen, Z.; Ren, Z.; Lian, T.; Ma, J. Neural Attentive Session-based Recommendation. In Proceedings of the CIKM ’17: 2017 ACM on Conference on Information and Knowledge Management, Singapore, 6–10 November 2017; pp. 1419–1428. [Google Scholar] [CrossRef]
  28. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention Is All You Need. In Advances in Neural Information Processing Systems (NeurIPS 2017); Curran Associates Inc.: Red Hook, NY, USA, 2017; Volume 30. [Google Scholar]
  29. Kang, W.C.; McAuley, J. Self-Attentive Sequential Recommendation. In Proceedings of the 2018 IEEE International Conference on Data Mining (ICDM), Singapore, 17–20 November 2018; pp. 197–206. [Google Scholar] [CrossRef]
  30. Sun, F.; Liu, J.; Wu, J.; Pei, C.; Lin, X.; Ou, W.; Jiang, P. BERT4Rec: Sequential Recommendation with Bidirectional Encoder Representations from Transformer. In Proceedings of the CIKM ’19: 28th ACM International Conference on Information and Knowledge Management, Beijing, China, 3–7 November 2019; pp. 1441–1450. [Google Scholar] [CrossRef]
  31. Chen, J.; Liu, Z.; Huang, X.; Wu, C.; Liu, Q.; Jiang, G.; Pu, Y.; Lei, Y.; Chen, X.; Wang, X.; et al. When Large Language Models Meet Personalization: Perspectives of Challenges and Opportunities. World Wide Web 2024, 27, 42. [Google Scholar] [CrossRef]
  32. Lyu, H.; Jiang, S.; Zeng, H.; Xia, Y.; Wang, Q.; Zhang, S.; Chen, R.; Leung, C.; Tang, J.; Luo, J. LLM-Rec: Personalized Recommendation via Prompting Large Language Models. In Proceedings of the Findings of the Association for Computational Linguistics: NAACL 2024, Mexico City, Mexico, 16–21 June 2024; pp. 583–612. [Google Scholar] [CrossRef]
  33. Tan, Z.; Liu, Z.; Jiang, M. Personalized Pieces: Efficient Personalized Large Language Models through Collaborative Efforts. arXiv 2024, arXiv:2406.10471. [Google Scholar] [CrossRef]
  34. Boz, A.; Zorgdrager, W.; Kotti, Z.; Harte, J.; Louridas, P.; Karakoidas, V.; Jannach, D.; Fragkoulis, M. Improving Sequential Recommendations with LLMs. ACM Trans. Recomm. Syst. 2025, 4, 19. [Google Scholar] [CrossRef]
  35. Wang, S.; Fan, W.; Feng, Y.; Shanru, L.; Ma, X.; Wang, S.; Yin, D. Knowledge graph retrieval-augmented generation for llm-based recommendation. In Proceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), Vienna, Austria, 27 July–1 August 2025; pp. 27152–27168. [Google Scholar]
  36. Salemi, A.; Kallumadi, S.; Zamani, H. Optimization Methods for Personalizing Large Language Models through Retrieval Augmentation. In Proceedings of the 47th International ACM SIGIR Conference on Research and Development in Information Retrieval, Washington, DC, USA, 14–18 July 2024; pp. 752–762. [Google Scholar] [CrossRef]
  37. Wang, Z.; Yu, Y.; Zheng, W.; Ma, W.; Zhang, M. MACRec: A Multi-Agent Collaboration Framework for Recommendation. In Proceedings of the 47th International ACM SIGIR Conference on Research and Development in Information Retrieval (SIGIR 2024), Washington, DC, USA, 14–18 July 2024; pp. 2760–2764. [Google Scholar] [CrossRef]
  38. Zhang, J.; Bao, K.; Wang, W.; Zhang, Y.; Shi, W.; Xu, W.; Feng, F.; Chua, T.S. Envisioning Recommendations on an LLM-Based Agent Platform. Commun. ACM 2025, 68, 48–57. [Google Scholar] [CrossRef]
  39. Guo, D.; Zhang, J.; Yang, B.; Lin, Y. Multi-modal intelligent situation awareness in real-time air traffic control: Control intent understanding and flight trajectory prediction. Chin. J. Aeronaut. 2025, 38, 103376. [Google Scholar] [CrossRef]
  40. Hebert, L.; Sayana, K.; Jash, A.; Karatzoglou, A.; Sodhi, S.; Doddapaneni, S.; Cai, Y.; Kuzmin, D. PERSOMA: PERsonalized SOft ProMpt Adapter Architecture for Personalized Language Prompting. arXiv 2024, arXiv:2408.00960. [Google Scholar] [CrossRef]
  41. Kojima, T.; Gu, S.S.; Reid, M.; Matsuo, Y.; Iwasawa, Y. Large Language Models Are Zero-Shot Reasoners. In Proceedings of the 36th International Conference on Neural Information Processing Systems (NeurIPS 2022), New Orleans, LA, USA, 28 November–9 December 2022; pp. 1613–1627. [Google Scholar]
  42. Salemi, A.; Mysore, S.; Bendersky, M.; Zamani, H. Lamp: When large language models meet personalization. arXiv 2023, arXiv:2304.11406. [Google Scholar]
  43. Sun, Z.; Liu, H.; Qu, X.; Feng, K.; Wang, Y.; Ong, Y.S. Large Language Models for Intent-Driven Session Recommendations. In Proceedings of the 47th International ACM SIGIR Conference on Research and Development in Information Retrieval, Washington, DC, USA, 14–18 July 2024; Association for Computing Machinery: New York, NY, USA, 2024; pp. 324–334. [Google Scholar] [CrossRef]
  44. Kuzilek, J.; Hlosta, M.; Zdrahal, Z. Open University Learning Analytics dataset. Sci. Data 2017, 4, 170171. [Google Scholar] [CrossRef]
  45. Zhuang, Y.; Kim, J. A BERT-Based Multi-Criteria Recommender System for Hotel Promotion Management. Sustainability 2021, 13, 8039. [Google Scholar] [CrossRef]
  46. Kille, B.; Lommatzsch, A. Defining a Meaningful Baseline for News Recommender Systems. In Proceedings of the INRA@ RecSys, Copenhagen, Denmark, 20 September 2019; pp. 24–28. [Google Scholar]
  47. Office of the Australian Information Commissioner. The Privacy Act. 2025. Available online: https://www.oaic.gov.au/privacy/privacy-legislation/the-privacy-act (accessed on 28 March 2025).
  48. Lim, S.; Oh, J. A Global Comparative Analysis of Data Protection Laws. IET Inf. Secur. 2025, 2025, 5536763. [Google Scholar] [CrossRef]
Figure 1. The zero-shot-to-head-shot hyperpersonalization framework illustrates the staged progression of hyperpersonalization along three stages spanning across four conceptual layers. It demonstrates how hyperpersonalization systems evolve incrementally from using limited prior data to building intent-aware user profiles through adaptive modules and GenAI techniques, enabling real-time personalization.
Figure 1. The zero-shot-to-head-shot hyperpersonalization framework illustrates the staged progression of hyperpersonalization along three stages spanning across four conceptual layers. It demonstrates how hyperpersonalization systems evolve incrementally from using limited prior data to building intent-aware user profiles through adaptive modules and GenAI techniques, enabling real-time personalization.
Asi 08 00186 g001
Figure 2. The ZDMM maps explicit, implicit, passive, and derived data modalities to each stage, illustrating how data quality and contextual depth increase over time to support hyperpersonalization.
Figure 2. The ZDMM maps explicit, implicit, passive, and derived data modalities to each stage, illustrating how data quality and contextual depth increase over time to support hyperpersonalization.
Asi 08 00186 g002
Figure 3. Use case illustration of how the Z2H2 framework progresses from zero-shot to few-shot and head-shot phases in an online education platform. The zero-shot phase uses limited learner metadata for initial personalization through prompting. The few-shot phase incorporates behavioral data to refine the BP and adapt recommendations dynamically. In the head-shot phase, the platform constructs a CIP, integrates real-time knowledge via RAG, and applies agentic AI for reasoning, tool calling, and targeted delivery to achieve precise, emotionally aligned hyperpersonalization.
Figure 3. Use case illustration of how the Z2H2 framework progresses from zero-shot to few-shot and head-shot phases in an online education platform. The zero-shot phase uses limited learner metadata for initial personalization through prompting. The few-shot phase incorporates behavioral data to refine the BP and adapt recommendations dynamically. In the head-shot phase, the platform constructs a CIP, integrates real-time knowledge via RAG, and applies agentic AI for reasoning, tool calling, and targeted delivery to achieve precise, emotionally aligned hyperpersonalization.
Asi 08 00186 g003
Figure 4. Sensitivity analysis for the few-shot and head-shot stages, showing how HR@5 varies with different smoothing and fusion parameters. (a) HR@5 as a function of α for different values of γ in the few-shot stage. (b) HR@5 as a function of β for different values of λ in the head-shot stage.
Figure 4. Sensitivity analysis for the few-shot and head-shot stages, showing how HR@5 varies with different smoothing and fusion parameters. (a) HR@5 as a function of α for different values of γ in the few-shot stage. (b) HR@5 as a function of β for different values of λ in the head-shot stage.
Asi 08 00186 g004
Table 1. Categorization of data modalities for hyperpersonalization.
Table 1. Categorization of data modalities for hyperpersonalization.
ModalityDescriptionFocus Area
ExplicitData directly provided by the user through intentional input mechanisms.Identity and Profile
ImplicitBehavioral data captured during human–system interaction, inferred without explicit input.Behavior and Usage
PassiveData obtained from the system, user environment, or situational context, independent of active user input.Context and Environment
DerivedData inferred or computed using machine learning models, GenAI, or aggregated external knowledge.Intent and Cognitive Inference
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Dandeniya, K.; Saltis, S.; Jayatilleke, S.; Mills, N.; Moraliyage, H.; De Silva, D.; Manic, M. Zero-Shot to Head-Shot: Hyperpersonalization in the Age of Generative AI. Appl. Syst. Innov. 2025, 8, 186. https://doi.org/10.3390/asi8060186

AMA Style

Dandeniya K, Saltis S, Jayatilleke S, Mills N, Moraliyage H, De Silva D, Manic M. Zero-Shot to Head-Shot: Hyperpersonalization in the Age of Generative AI. Applied System Innovation. 2025; 8(6):186. https://doi.org/10.3390/asi8060186

Chicago/Turabian Style

Dandeniya, Kanishka, Sam Saltis, Shalinka Jayatilleke, Nishan Mills, Harsha Moraliyage, Daswin De Silva, and Milos Manic. 2025. "Zero-Shot to Head-Shot: Hyperpersonalization in the Age of Generative AI" Applied System Innovation 8, no. 6: 186. https://doi.org/10.3390/asi8060186

APA Style

Dandeniya, K., Saltis, S., Jayatilleke, S., Mills, N., Moraliyage, H., De Silva, D., & Manic, M. (2025). Zero-Shot to Head-Shot: Hyperpersonalization in the Age of Generative AI. Applied System Innovation, 8(6), 186. https://doi.org/10.3390/asi8060186

Article Metrics

Back to TopTop