Abstract
The scientific literature is expanding at an unprecedented pace, making it increasingly difficult for researchers, students, and professionals to extract relevant insights efficiently. Traditional academic publishing offers static, one-size-fits-all content that does not cater to the diverse backgrounds, expertise levels, and interests of readers. This paper explores how generative AI can dynamically personalize scholarly content by tailoring summaries and key takeaways to individual user profiles. Nine scientific articles from a single journal issue were used to create the dataset, and prompt engineering was applied to generate tailored insights for exemplary personas: a digital humanities and open science researcher, and a mining and raw materials industry specialist. The effectiveness of AI-generated content modifications in enhancing readability, comprehension, and relevance was evaluated. The results indicate that generative AI can successfully emphasize different aspects of an article, making it more accessible and engaging to specific audiences. However, challenges such as content oversimplification, potential biases, and ethical considerations remain. The implications of AI-powered personalization in scholarly communication are discussed, and future research directions are proposed to refine and optimize AI-driven adaptive reading experiences.
Keywords:
GenAI-driven content personalization in academia; recommender systems for scientific literature; generative AI in scholarly publishing; Large Language Models (LLMs) in academic research; GenAI-assisted scientific text summarization; scientific content customization using AI; personalized academic reading with AI; adaptive summaries for scholarly communication; AI-driven knowledge dissemination in science; adaptive scholarly article retrieval through AI 1. Introduction
The rapid expansion of scientific knowledge has led to an overwhelming influx of academic studies Bornmann and Mutz (2015); Hanson et al. (2024); Jin (2024), making it increasingly difficult for researchers, students, and professionals to extract relevant information efficiently Santini (2024). The volume of published research articles continues to grow exponentially, with millions of new papers added to digital repositories each year across various disciplines (e.g., Boboris (2023), arXiv (2021)). As scientific inquiry becomes more complex and interdisciplinary Nguyen (2024), vast amounts of information must be sifted through to identify key insights that are most relevant to specific expertise, interests, or research objectives. While technological advancements have facilitated access to the scientific literature through digital databases and search engines Pranckutė (2021), the fundamental structure of academic publishing remains largely static Santini (2024). Research articles are typically written in a standardized format that does not account for the diverse needs of different readers.
While traditional methods like keyword searches and recommendation systems are valuable for identifying potentially relevant articles, they fall short in truly adapting the content to individual reader needs Rosman et al. (2016). These tools primarily focus on directing users to existing information but lack the capacity to modify the presentation, level of detail, or focus of a text to suit different backgrounds or research goals. For instance, a student encountering a paper outside their immediate expertise might miss crucial foundational concepts buried within technical details, even if the topic broadly aligns with their interests. Similarly, a practitioner seeking practical applications might struggle to extract actionable insights from a research paper primarily focused on theoretical contributions. The novelty of the approach presented in this paper lies in its potential to overcome these limitations by leveraging generative AI to dynamically tailor the content of scientific articles, thereby offering a more personalized and effective reading experience.
Recent advancements in artificial intelligence, particularly in the field of generative models, have opened new possibilities for enhancing the accessibility and personalization Chen et al. (2024) of scientific content. Large Language Models such as GPT-4, Claude, and Gemini demonstrate an unprecedented ability to generate Wu (2024), summarize Pu et al. (2023), and reformat textual content based on user input Chi et al. (2023). By leveraging prompt engineering Marvin et al. (2023), AI systems can tailor information to align with the background, knowledge level, and preferences of different readers. Such dynamic adaptation has the potential to revolutionize how scientific information is consumed, allowing users to engage with research articles in a manner that optimally suits their cognitive needs and expertise. In this study, the application of generative AI in customizing scholarly content to enhance reader engagement is explored. Specifically, modifications in the presentation of scientific articles are examined to determine whether AI can adjust content based on the distinct profiles of a researcher from academia and a specialist from industry. By generating personalized abstracts and content highlights, AI could enable more efficient access to information, improving comprehension and retention while reducing the time required to extract relevant insights. However, this approach also presents several challenges, including risks of information oversimplification, bias in content selection, and ethical concerns related to AI-generated modifications in scientific communication.
The primary objective of this study was to demonstrate how generative AI can personalize the reading experience of the scientific literature by dynamically adapting content to different user profiles. The hypothesis is that AI-driven customization can improve the accessibility and relevance of scholarly articles for diverse audiences, making scientific knowledge dissemination more efficient and inclusive. To assess the effectiveness, advantages, and limitations of this approach in academic publishing, an overview of current methods for content personalization in scientific publishing is first provided, highlighting existing limitations and identifying the gap that generative AI seeks to address. An experimental framework is then introduced, describing the selection of scientific articles, the creation of user personas, and the application of AI-generated prompts to customize content. The results of these experiments are analyzed to evaluate the effectiveness of AI-driven modifications in enhancing readability and user engagement. Finally, the broader implications of AI-driven personalization in scholarly communication are discussed, along with potential ethical concerns and directions for future research. By addressing these topics, this study contributes to the ongoing discourse on AI applications in academia, offering insights into how generative models can reshape the way scientific knowledge is disseminated and consumed.
2. Background
The increasing volume of the scientific literature has necessitated the development of tools that help researchers navigate and process vast amounts of information efficiently Alvarez et al. (2020); Knoth et al. (2023); Shin et al. (2019). Several approaches to content personalization in academic publishing have been explored, primarily through recommendation systems and automated summarization techniques. Recommendation algorithms, commonly used in digital libraries and academic search engines, suggest relevant articles based on citation networks, keyword similarity, or user preferences. Platforms such as Google Scholar, Semantic Scholar, ResearchGate, and Arxiv employ these methods to assist users in discovering research aligned with their interests. However, while these systems enhance content discovery, they do not alter the way information is presented, leaving the cognitive burden of comprehension entirely on the reader. Automated summarization tools represent another approach Zakkas et al. (2024), where AI-driven models extract key sentences from texts to generate abstracts or highlights. Such techniques, although useful, often produce generic outputs that do not account for variations in reader expertise or specific informational needs. The fundamental limitation of these traditional approaches is their inability to deeply adapt content presentation to different audiences. Scientific articles remain static, forcing readers to engage with the same level of technical detail regardless of their background knowledge, reading preferences, or research objectives. This lack of flexibility creates barriers to efficient knowledge acquisition, particularly for interdisciplinary researchers, students, and professionals seeking to grasp complex topics outside their immediate field of expertise.
The emergence of generative AI has introduced new possibilities for overcoming these limitations by enabling dynamic content adaptation. Large Language Models (LLMs) such as GPT-4, Claude, and Gemini operate using advanced natural language processing techniques, allowing them to generate human-like text Zhao et al. (2023), summarize information Pu et al. (2023), and rephrase content based on contextual prompts Chi et al. (2023). These models process vast corpora of textual data Villalobos et al. (2022), learning linguistic patterns, domain-specific terminology, and argument structures, which enables them to reformat information in ways that align with user needs. In the context of academia, generative AI has already been applied in tasks such as automated translation, trend analysis in scientific publishing, and summarization of research papers. AI-powered writing assistants help scholars draft manuscripts King (2023), refine abstracts, and generate literature reviews Antu et al. (2023), improving the efficiency of academic writing Bom (2023) and editing processes Shmueli et al. (2023). Moreover, some AI-driven tools provide interactive explanations of scientific concepts, offering simplified or in-depth versions of content depending on user input. These developments indicate that AI has the potential to facilitate personalized engagement with the scholarly literature by tailoring the way information is presented.
While a comprehensive exploration of LLM-driven personalization methods for scientific texts remains relatively underexplored in the current academic discourse, a pertinent analogy can be drawn to e-commerce recommendation systems, where the “buyers” are represented by scientists, researchers, students, engineers, and industry professionals, and the “products” are the formally structured scientific communications found in journal articles; acknowledging this parallel justifies a review of relevant research in the e-commerce domain, though it is crucial to recognize that the “product” in our context presents a significantly higher degree of semantic and syntactic complexity compared to, for example, typical household goods, thereby imposing considerably more intricate demands on LLM technologies to achieve a valuable representation of scientific work, necessitating the development of dedicated approaches and methodologies. However, the latter sector, driven by its considerably larger scale and direct commercial implications, has witnessed a more rapid and extensive adoption of advanced personalization techniques. Nevertheless, recent research has increasingly explored the potential of Large Language Models to revolutionize recommendation systems, moving beyond traditional methods. A comprehensive survey Wang et al. (2024) highlights this paradigm shift, emphasizing the remarkable language understanding, generalization capabilities, and reasoning skills of LLMs and their potential to significantly enhance recommendation tasks from the perspective of the recommender system community. This is particularly relevant as traditional deep neural network-based recommender systems often exhibit limitations in effectively capturing textual side information, generalizing across diverse scenarios, and providing reasoning for their predictions, as noted in another survey Fan (2024). One promising application of LLMs in this domain is the generation of high-quality explanations for recommended items Lubos et al. (2024), which can significantly improve user trust and decision-making. Furthermore, advancements have demonstrated how leveraging the reasoning capabilities of LLMs, particularly through techniques like chain-of-thought prompting, can enhance personalized recommender systems by better capturing subjective user preferences, an area previously underexplored in LLM reasoning applications Tsai et al. (2024). These advancements in harnessing LLMs for recommendation systems underscore the growing interest and potential in leveraging these powerful models for sophisticated information personalization tasks, providing a relevant backdrop for the exploration of their application in the context of the scientific literature.
Despite these advancements, there is a notable research gap in the application of generative AI for real-time, reader-specific customization of scientific content. Existing AI applications primarily focus on summarization and search optimization Kreutz and Schenkel (2022); Zhang et al. (2023) rather than dynamically adjusting content presentation to accommodate varying levels of expertise and reader preferences. The ability to generate tailored abstracts, reframe technical discussions, or emphasize specific aspects of a paper based on user needs remains largely unexplored. To address this gap, this study proposes the use of prompt engineering as a method for leveraging generative AI to modify scientific texts according to distinct reader personas. By defining specific instructions for AI models, it becomes possible to guide content adaptation, ensuring that different audiences receive information in a format that is most relevant and accessible to them. This approach aims to enhance the efficiency of academic reading, lower barriers to interdisciplinary knowledge transfer, and improve overall engagement with the scientific literature.
3. Materials and Methods
The methodology employed in this study, the flow of which is visually represented in Figure 1, was designed to assess the ability of AI-generated content to adapt to the needs of diverse academic and professional users. This section provides a comprehensive overview of the process, beginning with a selection of scientific articles, followed by the creation of distinct user personas, the definition of AI prompts tailored to their expectations, and the implementation of AI models for content generation.
Figure 1.
Research design for evaluating AI in content personalization.
3.1. Selection of Scientific Articles
To ensure a diverse and representative dataset for testing AI-generated content, a selection of nine academic articles from Publications, Volume 12, Issue 1 (March 2024) Publications, Volume 12, Issue 1 (2024), was chosen as the primary corpus for analysis. This issue covers a broad range of topics within the domain of scholarly communication, research evaluation, open science, and bibliometric trends, providing a suitable basis for assessing how AI adapts content to different audiences. The selected articles examine issues such as the role of ChatGPT in social sciences, the impact of citizen science initiatives on libraries, debates surrounding predatory journals, the application of FAIR principles in the humanities, and the evolution of bibliometric indicators in various research fields. Each article was carefully reviewed to extract key methodological insights and innovative aspects that would form the foundation for AI-generated content. The diversity of these articles ensured that AI-generated content would need to cater to different disciplinary perspectives and methodological approaches, thereby testing its ability to adjust content to distinct user expectations.
The selection of these articles was, in essence, made without prior testing on other collections to specifically identify a set where the demonstrated method would appear particularly effective. Instead, the decision was made to include all articles from a single special issue. This approach was adopted to provide a degree of informal confirmation that the choice of these specific articles did not significantly skew the results. The selection of a special issue with a relatively small number of articles represented a conscious trade-off. The aim was to present readers with the full texts of the results generated by the LLM, allowing for their independent assessment of the method’s effectiveness, while also ensuring the overall manuscript remained of a manageable length and did not become overly cumbersome.
Below is a brief overview of each article:
- Bibliometric Overview of ChatGPT: New Perspectives in Social SciencesThis study conducted a bibliometric analysis of ChatGPT’s impact on social sciences using Scopus data. It identified trends, co-citations, and knowledge gaps, emphasizing AI’s role in academic discourse.
- Benefits of Citizen Science for LibrariesThe article examines the role of citizen science in enhancing library functions. It systematically reviews the literature to outline how libraries can leverage citizen science to promote research engagement.
- Should I Buy the Current Narrative about Predatory Journals? Facts and Insights from the Brazilian ScenarioThis paper challenges prevailing assumptions about predatory journals, advocating for a nuanced debate on publication practices, impact factors, and the evolving landscape of academic publishing.
- FAIRness of Research Data in the European Humanities LandscapeThis article analyzes research data in the humanities, evaluating its openness, compliance with FAIR principles, and representation in repositories. It highlights challenges in accessibility and reusability.
- Reducing the Matthew Effect on Journal Citations through an Inclusive Indexing Logic: The Brazilian Spell ExperienceThis study explores how inclusive indexing in local databases can mitigate the Matthew effect in academic citations, fostering more equitable visibility of journals.
- Does Quality Matter? Quality Assurance in Research for the Chilean Higher Education SystemThe research assesses quality assurance in Chilean universities, revealing that accreditation mainly correlates with publication quantity rather than impact or quality.
- Mining and Mineral Processing Journals in the WoS and Their Rankings When Merging SCIEx and ESCI DatabasesThis article analyzes how merging SCIEx and ESCI databases in JCR 2022 affected journal rankings in the mining and mineral processing field, offering insights for researchers in the industry.
- Tracing the Evolution of Reviews and Research Articles in the Biomedical Literature: A Multi-Dimensional Analysis of AbstractsUsing computational linguistic analysis, this study examines shifts in the writing styles of biomedical research articles and reviews of over three decades.
- Going Open Access: The Attitudes and Actions of Scientific Journal Editors in ChinaThis study investigates Chinese journal editors’ perspectives on open access publishing, analyzing their motivations, barriers, and responses to academic publishing reforms.
3.2. User Persona Development
To assess the effectiveness of AI-generated personalized content, two distinct user personas were created, representing different academic and professional backgrounds.
- Persona A:Dr. Agnieszka Nowak—Digital Humanities and Open Science Researcher
- Occupation: Associate Professor at a university, specializing in digital humanities and social sciences;
- Interests: Open research data, FAIR principles, bibliometrics, open science, ethics of scientific publishing;
- What she looks for in the academic literature? She wants to understand how open science and data accessibility impact humanities and social sciences research. She is also interested in the role of AI (e.g., ChatGPT) in academia and education.
- Persona B:Eng. José Antonio García—Mining and Raw Materials Industry Specialist
- Occupation: Engineer specializing in mining and mineral processing, working for an industrial engineering company;
- Interests: Innovations in the extractive industry, trends in scientific publishing for technical fields, the impact of journal indexing on industry recognition;
- What he looks for in the academic literature?He seeks practical insights into scientific publishing trends in his field, as well as how indexing and citation metrics affect the recognition of technical research.
3.3. Definition of AI Prompts
To explore the potential of Large Language Models in personalizing academic content, a set of structured AI prompts was developed to adapt article summaries to different user personas. The personalization process followed five distinct strategies, each aimed at enhancing accessibility, relevance, and engagement with academic materials.
Each prompt included a common structure, where only the persona description varied. The four core personalization methods applied in the prompts were as follows:
- Highlighting original fragments of titles and abstracts that are particularly relevant to a given persona, using bold formatting to emphasize crucial aspects.
- Structuring abstracts into bullet-point lists to clearly delineate key research contributions, methodologies, and findings aligned with the persona’s interests.
- Ranking articles based on their relevance to the persona’s expertise, providing a rating with a justification.
- Generating personalized recommendations in the persona’s native language, explaining the article’s relevance and value in their specific research or professional context.
By embedding these strategies within structured prompts, it was ensured that the generated content was not only accurate but also aligned with the specific needs and expectations of different professional and academic users.
Each prompt was applied across different personas to ensure that the content was not only factually accurate but also contextually aligned with each user’s domain of expertise.
3.4. Implementation of AI
The personalization of academic content was carried out using the ChatGPT-4o language model, selected for its advanced text generation capabilities, strong contextual awareness, and adaptability in restructuring complex information. The implementation process involved several key steps to ensure the generated content was both accurate and contextually relevant to different personas. First, persona descriptions were standardized to maintain consistency across all applied prompts. Each persona was characterized by specific research interests, a professional background, and language preferences, allowing for tailored adjustments in the way academic abstracts were processed and presented.
Once the persona framework was established, the AI was fed with original academic abstracts, and each of the structured prompts was applied systematically. These prompts instructed GPT-4o to modify the abstracts in distinct ways, including highlighting critical phrases, simplifying language, structuring content in a more digestible format, ranking articles based on relevance, and generating personalized recommendations in the persona’s native language.
This task was accomplished using the ChatGPT-4o model with the “reason” (think before responding) option enabled, facilitating a more thorough analysis and consideration of each step in the personalization process.
The execution of each prompt was conducted using the standard chat interface of ChatGPT-4o without leveraging any external API services. Each prompt was initiated within a newly instantiated chat session. This methodological choice ensured that the outcomes of each prompt were not influenced by the contextual history of prior conversational turns. Such an approach was adopted to enhance the transparency and replicability of the research. It is important to acknowledge that Large Language Models operate based on probabilistic distributions, and the inherent stochasticity, due to the lack of control over the random number generator seed, implies that the results obtained in subsequent experiments may not be perfectly identical. However, empirical observations suggest that the outputs remain highly consistent across multiple executions.
4. Results
To evaluate the capability of Large Language Models in tailoring scientific information to specific user profiles, a structured prompt (Listing 1) was employed to analyze a selection of nine distinct scientific articles. This prompt was designed to simulate a scenario where an LLM acts as an information filter, highlighting sections of interest within the article’s title, keywords, and abstract for two predefined personas, Persona A and Persona B.
For each of the nine articles, the prompt was executed once, encompassing the analysis for both personas within a single invocation. This process involved providing the LLM with the article’s title, keywords, and abstract, along with detailed descriptions of Persona A and Persona B, outlining their respective interests and areas of focus. The LLM was then instructed to identify and emphasize (using bold formatting) the portions of the title, keywords, and abstract that would be most relevant and engaging for each persona.
This methodology allowed for the exploration of the potential of LLMs in personalizing scientific information retrieval, demonstrating their capacity to adapt to diverse user needs and preferences. By showcasing the LLM’s ability to filter and prioritize information based on persona-specific criteria, the aim was to highlight their potential in enhancing the efficiency and relevance of scientific literature descriptions.
The results obtained from the application of the persona-specific prompt were compiled and are presented in Table 1 (for the first five articles) and Table 2 (for the subsequent four articles). These tables provide a detailed overview of the LLM’s performance in tailoring scientific information to the defined user personas. Each table is structured with two columns: the first column displays the personalized titles, keywords, and abstracts for Persona A, while the second column presents the corresponding personalized outputs for Persona B, allowing for a direct comparison of the LLM’s ability to cater to distinct user profiles.
Table 1.
Personalized scientific article elements (titles, keywords, abstracts) generated by the LLM for Personas A and B (Articles 1–5).
Table 2.
Personalized scientific article elements (titles, keywords, abstracts) generated by the LLM for Personas A and B (Articles 6–9).
Next, a single prompt was constructed to assess the LLM’s capability to estimate the potential interest of two distinct personas across a set of nine scientific articles. This approach allowed for the simultaneous injection of all nine articles (Listing 2) into the LLM, streamlining the evaluation process. The prompt was meticulously designed to instruct the LLM to provide a Likert scale rating (1–5) for each article and each persona, accompanied by a detailed justification for each rating. The justifications were required to explain the rationale behind the assigned interest level based on the title and abstract of each article and in consideration of the specific characteristics of Persona A and Persona B.
| Listing 1. The prompt used to generate the persona-targeted analysis of a scientific article’s title, keywords, and abstract. |
| I need you to analyze a scientific article and highlight sections of interest for two distinct personas, Persona A and Persona B. For each persona, please review the provided title, keywords, and abstract of the scientific article. Identify the parts of the title, keywords, and abstract that would be most relevant and engaging for each persona based on their characteristics. Then, for both Persona A and Persona B, present the title, keywords, and abstract of the article, with the sections of interest bolded. Maintain the original text of the title, keywords, and abstract, only adding bold formatting to emphasize the relevant parts for each persona. Please provide two outputs: one for Persona A and one for Persona B, each containing the title, keywords, and abstract with the relevant sections bolded. Persona A: (…) Persona B: (…) Scientific article: Title: (…) Keywords: (…) Abstract: (…) |
Upon execution, the LLM systematically processed each article, generating a comprehensive output that included the assigned Likert scale rating and a corresponding justification for both personas. The output was structured to clearly indicate the persona, the article title, the Likert scale rating, and the justification for each rating. This format facilitated a clear and organized presentation of the LLM’s assessments, enabling a thorough analysis of its performance in estimating persona-specific interest in the scientific literature.
The results of this analysis are presented in Table 3, which provides a comprehensive overview of the LLM’s estimations. The table is structured with three columns: the first column lists the titles of the analyzed scientific articles; the second column displays the Likert scale rating and justification provided by the LLM from the perspective of Persona A; and the third column presents the corresponding rating and justification for Persona B. This tabular format allows for a clear and direct comparison of the LLM’s assessments across the two personas and the nine articles, facilitating a thorough examination of the LLM’s performance in this task.
Table 3.
Likert scale ratings and justifications provided by the LLM for Persona A and Persona B across nine scientific articles.
| Listing 2. The prompt employed to gauge potential interest in scientific articles using an LLM and Likert scale ratings. |
| I need you to estimate the potential level of interest for two distinct personas, Persona A and Persona B, in a set of nine scientific articles. For each persona and each of the nine scientific articles (described below by their title and abstract), please provide a rating on a Likert scale indicating the potential level of interest. Likert Scale: 1 – Not at all interested 2 – Slightly interested 3 – Moderately interested 4 – Very interested 5 – Extremely interested For each of the nine articles and for both Persona A and Persona B, please provide a Likert scale rating (1–5) along with a brief justification for your rating. The justification should explain why you believe that persona would have that level of interest, based on the title, and abstract of the article. Please present your output clearly, indicating the persona, the article number (1–9), the Likert scale rating, and the justification for each rating. Remember to consider the characteristics of Persona A and Persona B when evaluating the relevance and appeal of each article’s title and abstract. Persona A: (…) Persona B: (…) Scientific articles: (…) |
Finally, in a further demonstration of the LLM’s capability to process and tailor information, a consolidated prompt was utilized, incorporating all nine scientific articles (Listing 3) simultaneously. This approach aimed to assess whether the LLM could efficiently manage a larger volume of input while maintaining the precision and relevance of its outputs. The prompt, structured to elicit persona-specific key points and their translated equivalents, was executed in a single iteration.
Upon execution, the LLM systematically processed each article, extracting the core informational elements as defined by their titles and abstracts. Subsequently, these key points were meticulously adapted to align with the distinct interests and backgrounds of the two specified personas. Crucially, the LLM demonstrated an ability to not only distill the essential content of each article but also to contextualize it within the cognitive framework of each persona. This was evidenced by the nuanced variations in the summaries, reflecting the personas’ differing perspectives and areas of focus.
Furthermore, the prompt’s design mandated the translation of each key point into the respective native languages of the personas. The LLM successfully executed this translation, providing bilingual outputs that retained the semantic integrity of the original English summaries. This capability underscores the LLM’s potential as a tool for cross-cultural communication and information dissemination, particularly in contexts where tailored information delivery is paramount.
The successful execution of this consolidated prompt highlights the LLM’s scalability and adaptability. By efficiently processing multiple documents and generating persona-specific summaries, the LLM showcased its potential to streamline information processing and delivery in various applications, including research, education, and personalized content generation.
The results obtained from the consolidated prompt were subsequently compiled and presented in tabular format. Specifically, Table 4 showcases the persona-specific summaries for the first five articles, while Table 5 details the summaries for the remaining four. Mirroring the structure of Table 3, each table is organized with the article title in the first column, followed by bilingual key points tailored to Persona A’s perspective in the second column, and finally, bilingual key points tailored to Persona B’s perspective in the third column. This consistent formatting facilitates a direct comparison of the LLM’s outputs across all articles and personas, allowing for a comprehensive analysis of the model’s ability to adapt and translate scientific information.
Table 4.
Persona-tailored article summaries (first five articles) with English and native-language key points, as generated by the LLM.
Table 5.
Persona-tailored article summaries (last four articles) with English and native-language key points, as generated by the LLM.
| Listing 3. Structured prompt for generating translated, persona-focused summaries of research articles. |
| For each of the nine articles, and for each of the two personas: 1. Identify Key Points: Analyze the article (based on its title and abstract) and determine up to three key points that summarize its main content. 2. Persona-Specific Relevance: Tailor these key points to be relevant and interesting to each persona, considering their described interests and background. Output Format: For each article, present the output in the following format: Article: [Article Title] Persona 1 Perspective: – (English) Point 1 (([Persona 1 Native Language]) Point 1 Translation) – (English) Point 2 (([Persona 1 Native Language]) Point 2 Translation) – (English) Point 3 (if applicable) (([Persona 1 Native Language]) Point 3 Translation (if applicable)) Persona 2 Perspective: – (English) Point 1 (([Persona 2 Native Language]) Point 1 Translation) – (English) Point 2 (([Persona 2 Native Language]) Point 2 Translation) – (English) Point 3 (if applicable) (([Persona 2 Native Language]) Point 3 Translation (if applicable)) Please provide the analysis in the structured format described above. Persona A: (…) Persona B: (…) Scientific articles: (…) |
5. Discussion
The assumption of having persona information for users searching bibliographic data is undoubtedly a significant consideration. However, it is important to note that persona profiles can be constructed through various means, both explicit and implicit. Firstly, during user registration, direct inquiries regarding interests can be employed to gather initial data. Secondly, supplementary information can be sourced from platforms such as ORCID, enriching the persona’s profile with professional and academic details. Finally, a dynamic persona profile can be progressively developed by analyzing user activity and tracked interests derived from search patterns and interactions with data. In this latter scenario, user registration is not a prerequisite, as identification can be facilitated through cookies, provided user consent is granted. This approach allows for the creation of nuanced and evolving persona profiles, enhancing the relevance of information retrieval.
With a sufficiently detailed user profile established, a range of personalization techniques can be employed to present bibliographic data in a manner that is more relevant and conducive to the user’s specific needs. This tailored presentation facilitates quicker, easier, and more informed decisions regarding the selection of pertinent bibliographic records. For instance, personalized recommendations, customized search results, and context-aware summaries can be generated to align with the user’s identified interests and research focus. By leveraging user persona data, the relevance and efficiency of bibliographic information retrieval can be significantly enhanced, empowering users to navigate the vast landscape of the scholarly literature with greater precision.
In this study, a corpus of nine articles underwent experimental enhancement using three distinct prompts. The first involved bold formatting of selected sections of the title, keywords, and abstract. The second prompt aimed to estimate the persona’s interest level, supported by detailed justifications. The third focused on identifying key points of interest for the persona, along with their translations. Detailed discussions on the effects of each of these three prompts were illustrated using four selected examples, which the author deemed worthy of more in-depth analysis. These four cases were chosen to effectively demonstrate the method’s efficacy in directing attention to the most relevant content for each target audience.
In the case of the article “Bibliometric Overview of ChatGPT: New Perspectives in Social Sciences”, the bold formatting for Persona A was designed to foreground elements that resonate with a digital humanities scholar. The phrase “Bibliometric Overview of ChatGPT” was emphasized in the title to highlight the focus on bibliometric analysis—a method that is integral to evaluating emerging digital tools in academia. Additionally, keywords such as “ChatGPT; artificial intelligence; bibliometric analysis; ethical implications; educational technology” were accentuated so that the intersections between AI applications, ethical considerations, and educational innovation were immediately visible. Conversely, for Persona B, the formatting was slightly adjusted; the emphasis was shifted toward “bibliometric analysis” in the keywords and on methodological aspects in the abstract (e.g., “co-citations, keywords and international collaborations”) to appeal to an audience with a practical interest in the analytical and indexing dimensions of scientific publishing.
The article “FAIRness of Research Data in the European Humanities Landscape” provides another illustrative example. For Persona A, who is invested in open science and digital humanities, the text was tailored to underscore concepts central to data openness and ethical research practices. Phrases in bold such as “FAIRness of Research Data”, “humanities”, and “openness” in the keywords, as well as multiple references to “FAIR principles” in the abstract, were intended to immediately signal the article’s relevance to the challenges of data sharing and the governance of humanities research data. For Persona B, although the overall content remained identical, the bold formatting was adapted to highlight technical aspects—such as the emphasis on “datasets” and “research data”—thereby aligning the presentation with an audience that values precision in data metrics and the infrastructural dimensions of research information management.
A particularly clear demonstration of targeted emphasis is provided by the article “Mining and Mineral Processing Journals in the WoS and Their Rankings When Merging SCIEx and ESCI Databases—Case Study Based on the JCR 2022 Data”. Given that Persona B’s professional interests were in the mining and extractive industries, the bold formatting in this instance was calibrated to highlight industry-specific and evaluative metrics. In both the title and the keywords, technical terms such as “WoS”, “SCIEx”, “ESCI”, and “JCR 2022 Data” were accentuated, thus foregrounding the methodological and bibliometric rigor that underpins journal indexing—a factor that is critical for assessing publication quality in technical domains. For Persona A, although similar bibliometric elements were marked, the degree of emphasis was moderated so as to balance the interdisciplinary appeal of the article while still acknowledging the significance of indexing systems in the broader context of scholarly communication.
Finally, the article “Going Open Access: The Attitudes and Actions of Scientific Journal Editors in China” was employed to illustrate how the method can be refined to cater to differing thematic priorities. For Persona A, who is highly engaged with open science, the bold formatting was concentrated on “Open Access” in the title and within the abstract, thereby accentuating the transformative potential of open access models in reshaping academic publishing practices. In contrast, for Persona B, the emphasis was realigned to underline the procedural and evaluative components of editorial practices. The bold formatting in this version accentuated “the Attitudes and Actions of Scientific Journal Editors” as well as key phrases concerning the mechanisms of academic publishing in China, which are aspects likely to be appreciated by an audience that is keenly attuned to industry trends and practical implications in scientific publishing.
Collectively, these four examples underscore that the use of prompt-driven bold formatting via a generative AI model can be effectively utilized to direct a reader’s attention toward those elements of a scientific text that are most likely to resonate with their specific academic and professional interests. The experimental results, thus, suggest that, through strategic textual emphasis, it is possible to personalize academic reading experiences in a manner that enhances both the accessibility and relevance of the scientific literature for diverse audiences.
The subsequent analysis is provided to illustrate, through four exemplary cases, how the proposed method was demonstrated to effectively capture divergent academic interests. In each instance, the generative model was prompted to evaluate the potential level of interest of two distinct personas—Persona A, a digital humanities and open science researcher, and Persona B, a mining and raw materials industry specialist—in the context of nine scientific articles. Four cases were selected that most prominently demonstrate the method’s ability to differentiate between the interests of these personas.
In the case of the “Bibliometric Overview of ChatGPT (…)”, the model’s evaluation revealed a pronounced divergence between the personas. Persona A was rated at the highest level (5-Extremely interested) due to the article’s focus on ChatGPT’s impact on social sciences and its bibliometric analysis. This directly aligned with her established interests in artificial intelligence, bibliometrics, and the digital transformation of research. In contrast, Persona B received a rating of 2 (slightly interested), as the thematic focus on social sciences and AI did not intersect significantly with his technical orientation and the priorities of the mining industry. This stark contrast demonstrates that the method is capable of aligning content relevance with the nuanced research profiles of different academic domains.
The evaluation of “FAIRness of Research Data in the European Humanities Landscape” further underscored the model’s sensitivity to disciplinary relevance. The article was rated 5 (extremely interested) for Persona A, whose research was deeply embedded in the exploration of FAIR principles and the challenges associated with data sharing in the humanities. The explicit focus on open science and data accessibility was deemed highly pertinent to her academic endeavors. Conversely, Persona B was assigned a rating of 1 (not at all interested), as the paper’s concentration on humanities research data and FAIR principles fell entirely outside his core technical and industry-centric interests. This case reinforces the method’s capacity to distinctly recognize and prioritize domain-specific content.
A particularly illustrative example of the method’s discriminative power is provided by the evaluation of the article “Mining and Mineral Processing Journals in the WoS and Their Rankings When Merging SCIEx and ESCI Databases (…)”. Here, the scoring was effectively inverted relative to the previous cases: Persona A was rated 1 (not at all interested) due to the article’s focus on mining and mineral processing—a field that is completely outside her realm of digital humanities and open science research—while Persona B was rated at 5 (extremely interested). The extreme divergence in ratings is indicative of the method’s robust capability to detect and differentiate between subject matter that is of paramount importance to one academic profile but entirely extraneous to another.
The final example is provided by the assessment of “Going Open Access: The Attitudes and Actions of Scientific Journal Editors in China”. In this instance, Persona A was once again rated 5 (extremely interested) because the article’s focus on open access practices and the editorial attitudes toward scientific publishing closely aligned with her primary research interests in open science. Persona B, while still recognizing the potential value of understanding publishing trends, was rated 3 (moderately interested) as the technical and industry-specific dimensions of his work rendered the topic only partially relevant. This nuanced differentiation underscores the method’s ability not only to segregate highly domain-specific content but also to capture subtleties in relative interest when the subject matter is tangentially pertinent to a persona’s focus.
It is thereby concluded that the experimental application of the Large Language Model with a structured prompt was demonstrated to successfully tailor academic content recommendations in accordance with distinct scholarly profiles. The four cases discussed herein reveal that the method is capable of producing finely tuned evaluations that mirror the varying priorities of the personas. In doing so, the approach offers significant potential for enhancing the personalization of academic reading experiences, ensuring that recommendations are more closely aligned with the precise needs and interests of diverse research communities. Such capability is poised to contribute substantially to the broader endeavor of tailoring scientific knowledge in a rapidly evolving academic landscape.
The final prompt presented in this research demonstrated the LLM’s capability to summarize key information and translate it into various foreign languages. Significantly, the translations into the personas’ native languages (Polish for Persona A and Spanish for Persona B) were inferred by the model despite their absence in the explicit persona descriptions, showcasing the method’s capacity for contextual inference.
The first example, Bibliometric Overview of ChatGPT: New Perspectives in Social Sciences, was chosen on the basis that it offered a dual-layered insight into both the bibliometric landscape of AI in social sciences and the ethical dimensions of digital transformation. It was observed that the generated key points were precisely aligned with the research interests of Persona A—who was engaged in digital humanities and open science—while also addressing the technical aspects pertinent to Persona B’s focus on research trends and citation metrics. In this instance, the ability of the system to translate the academic nuances into Polish and Spanish was found to be particularly compelling, as it illustrated the successful adaptation of scientific language to culturally relevant communicative forms.
The second example, Reducing the Matthew Effect on Journal Citations through an Inclusive Indexing Logic: The Brazilian Spell (Scientific Periodicals Electronic Library) Experience, was selected because it effectively conveyed a nuanced critique of prevailing citation biases and the implications of inclusive indexing practices. It was noted that the personalized summaries managed to articulate the ethical and methodological dimensions of the study in a manner that resonated with both personas. For Persona A, the emphasis was placed on the benefits of alternative metrics and open access paradigms in enhancing scholarly transparency, while for Persona B, practical insights into citation metrics and the operational aspects of journal indexing were foregrounded. This dual-contextualization served to confirm that the generative approach is capable of addressing multifaceted academic issues by generating outputs that are simultaneously precise and adaptable.
The third example, Mining and Mineral Processing Journals in the WoS and Their Rankings When Merging SCIEx and ESCI Databases—Case Study Based on the JCR 2022 Data, was selected for its capacity to capture the intricate dynamics associated with evolving journal ranking systems. In this case, the personalized key points were found to be effective in highlighting the technical challenges and changes induced by database mergers—a subject that was directly relevant to the industrial and engineering concerns of Persona B. At the same time, the summaries maintained an emphasis on transparency and the broader implications for scholarly communication, which were issues of considerable interest to Persona A. The coherent delivery of these dual perspectives provided further evidence of the method’s versatility and robustness.
The fourth example, Tracing the Evolution of Reviews and Research Articles in the Biomedical Literature: A Multi-Dimensional Analysis of Abstracts, was incorporated to demonstrate the method’s proficiency in addressing longitudinal research trends. In this instance, the study’s focus on the evolution of narrative structures and linguistic standardization over a span of three decades was distilled into key points that were both comprehensive and reflective of the specific informational needs of the target audiences. The generated summaries were observed to be not only faithful to the original analytical outcomes but also effectively rendered into the deduced native languages. This case thereby exemplified the method’s capability to adapt complex, multidimensional analyses into succinct, persona-tailored outputs.
In summary, it was substantiated that the experimental framework is capable of delivering personalized, dual-language summaries that are attuned to the distinct scholarly and professional orientations of diverse academic audiences. The selected examples illustrate the method’s capacity for contextual inference, linguistic adaptation, and domain-specific customization. It is anticipated that these findings will encourage further refinement and expansion of generative AI techniques in the realm of scientific communication, ultimately contributing to a more inclusive and accessible dissemination of academic knowledge.
The research presented herein is not without limitations, which are primarily inherent to newly proposed methodologies leveraging cutting-edge advancements in science and technology.
Firstly, it must be acknowledged that the reported findings are preliminary, with the promising initial results currently supported by the subjective analysis and evaluation of the author. Consequently, readers are encouraged to independently assess the extent to which these findings provide a foundation for future research endeavors. While the central focus of this paper lies in the clear exposition of the proposed method and the detailed presentation of results, exemplified by the inclusion of full LLM-generated texts, this approach necessitated a restriction on the number of scholarly works analyzed, representing a second limitation.
Furthermore, the scientific publications constituting the research data for this study were drawn from a single, albeit highly interdisciplinary, scientific domain. Future investigations should aim to validate the effectiveness of the proposed methodology across a broader spectrum of scientific disciplines, including those within STEM. Assessing the efficacy of this method, which fundamentally aligns with natural language processing techniques, presents a notable challenge for quantitative evaluations. Traditional text similarity metrics such as ROUGE and BLEU Graham (2015) may not prove particularly efficacious in this context. However, the technology underpinning the method itself, namely Large Language Models, may offer a solution. Specifically, embedding models could be employed to compute semantic vectors for individual personas, representing their interests, and for the personalized content generated by the LLM. Subsequently, the semantic distance between these vectors could be calculated for each persona. This approach would facilitate larger-scale studies and enable the automation of quality assessments for the generated content. Finally, it is pertinent to note that this study utilized state-of-the-art technology. Future research employing smaller, open-source language models could provide valuable insights into the feasibility of developing scientific content recommendation systems based on locally deployable and more cost-effective models, potentially offering enhanced privacy for processed information.
6. Conclusions
It can be concluded that the application of a Large Language Model for tailoring academic reading experiences was demonstrated to be effective in aligning scholarly content with the diverse interests of targeted personas. The experimental results showed that the proposed approach is capable of accentuating relevant aspects of academic articles—ranging from bibliometric analyses and ethical considerations to technical indexing practices—thereby enhancing both the accessibility and the contextual relevance of the scientific literature.
Furthermore, the ability of generative AI to personalize academic content holds significant potential for fostering and advancing interdisciplinary research. By tailoring summaries and key takeaways to the specific background and expertise of researchers from different fields, this method can help bridge the knowledge gaps that often hinders collaboration across disciplines. Researchers can more readily grasp the core concepts and relevance of work outside their primary domain, facilitating the cross-pollination of ideas and methodologies. This enhanced understanding can lead to the identification of novel connections between seemingly disparate fields, potentially sparking innovative research directions and solutions to complex, multifaceted problems that require interdisciplinary approaches. The ability to dynamically adapt scholarly content could, therefore, lower the barrier to entry for researchers seeking to explore the literature outside their immediate specialization, ultimately contributing to a more interconnected and collaborative scientific landscape.
For instance, consider a researcher in digital humanities interested in the environmental impact of technology. Using the proposed method, they could engage with a paper focused on the mining and raw materials industry. The AI could tailor the summary to highlight aspects related to resource depletion, waste management, and sustainable practices, making the technical details more accessible and relevant to their humanities perspective. Conversely, a mining engineer could use the same method to understand a paper from the field of open science that discusses data sharing and accessibility. The AI could emphasize the practical implications of open data principles for the mining industry, such as improved data analysis and collaboration. Another example could involve a scholar in bibliometrics using the method to understand a paper on ethical considerations in AI from a philosophical perspective. The AI could highlight the core ethical arguments and their potential relevance to the development and deployment of bibliometric indicators. These examples illustrate how the proposed method can act as a translator and contextualizer, enabling researchers from diverse disciplines to more effectively engage with and learn from each other’s work, ultimately fostering a more integrated and dynamic research environment.
Despite these promising findings, it must be acknowledged that the underlying LLM technology remains susceptible to hallucinations Huang et al. (2025) and biases Dai et al. (2024) that originate from the training datasets. It is, therefore, recommended that, in any practical implementation, a mechanism for human feedback be incorporated into the monitoring process. Even minimal feedback tools, such as like/dislike functionalities, are deemed essential to ensure that the generated outputs are critically evaluated and continuously improved.
Furthermore, future research is suggested to focus on the refinement of personalization algorithms, the integration of advanced bias detection and mitigation strategies, and the exploration of more comprehensive human-in-the-loop feedback mechanisms. These avenues are anticipated to further enhance the reliability and applicability of generative AI in the domain of scientific communication, ultimately contributing to a more inclusive and effective dissemination of academic knowledge.
To further advance the personalization and adaptation of scientific article information, several promising research directions can be identified. One key area is the incorporation of full-text analysis into the personalization framework. Moving beyond abstracts and selected segments, integrating the complete content of scientific articles could enable a more comprehensive extraction of nuanced key points and facilitate deeper semantic understanding, ultimately leading to richer, more tailored summaries.
Another important direction is the inclusion of bibliographic references within the personalized output. References and citation lists not only provide insights into the intellectual context and impact of a work but also offer a valuable resource for mapping scholarly networks. Future studies could investigate methods to integrate citation network analysis with content personalization, thereby allowing readers to better navigate the academic discourse and identify related works of interest.
Additionally, the development of advanced multi-modal models that synergistically combine full-text processing, bibliometric data, and contextual metadata should be pursued. Such models could harness the interplay between textual content and bibliographic structures to generate more refined, context-aware academic recommendations and summaries.
Funding
This research received no external funding.
Data Availability Statement
Data are contained within the article.
Conflicts of Interest
The author declares no conflicts of interest.
References
- Alvarez, E., Lamagna, F., Miquel, C., & Szewc, M. (2020). Intelligent arxiv: Sort daily papers by learning users topics preference. arXiv, arXiv:2002.02460. Available online: https://arxiv.org/abs/2002.02460 (accessed on 20 February 2025).
- Antu, S. A., Chen, H., & Richards, C. K. (2023, July 7). Using LLM (Large Language Model) to improve efficiency in literature review for undergraduate research. Workshop on Empowering Education with LLMs-the Next-Gen Interface and Content Generation 2023 co-located with 24th International Conference on Artificial Intelligence in Education (AIED 2023) (pp. 8–16), Tokyo, Japan. [Google Scholar]
- arXiv. (2021). arXiv statistics: 2021 submissions by area. Available online: https://info.arxiv.org/help/stats/2021_by_area/index.html (accessed on 20 February 2025).
- Boboris, K. (2023, November 3). arXiv sets new record for monthly submissions. Available online: https://blog.arxiv.org/2023/11/03/arxiv-sets-new-record-for-monthly-submissions/ (accessed on 20 February 2025).
- Bom, H.-S. H. (2023). Exploring the opportunities and challenges of ChatGPT in academic writing: A roundtable discussion. Nuclear Medicine and Molecular Imaging, 57(4), 165–167. [Google Scholar] [CrossRef] [PubMed]
- Bornmann, L., & Mutz, R. (2015). Growth rates of modern science: A bibliometric analysis based on the number of publications and cited references. Journal of the Association for Information Science and Technology, 66(11), 2215–2222. [Google Scholar]
- Chen, J., Liu, Z., Huang, X., Wu, C., Liu, Q., Jiang, G., Pu, Y., Lei, Y., Chen, X., Wang, X., Zheng, K., Lian, D., & Chen, E. (2024). When large language models meet personalization: Perspectives of challenges and opportunities. World Wide Web, 27(4), 42. [Google Scholar]
- Chi, A., Chen, L.-K., Chang, Y.-C., Lee, S.-H., & Chang, J. S. (2023). Learning to paraphrase sentences to different complexity levels. Transactions of the Association for Computational Linguistics, 11, 1332–1354. [Google Scholar] [CrossRef]
- Dai, S., Xu, C., Xu, S., Pang, L., Dong, Z., & Xu, J. (2024, August 25–29). Bias and unfairness in information retrieval systems: New challenges in the llm era. 30th ACM Sigkdd Conference on Knowledge Discovery and Data Mining (pp. 6437–6447), Barcelona, Spain. [Google Scholar]
- Fan, W. (2024). Recommender systems in the era of large language models (LLMs). IEEE Transactions on Knowledge and Data Engineering, 36(11), 6889–6907. [Google Scholar] [CrossRef]
- Graham, Y. (2015, September 17–21). Re-evaluating automatic summarization with BLEU and 192 shades of ROUGE. 2015 Conference on Empirical Methods in Natural Language Processing (pp. 128–137), Lisbon, Portugal. [Google Scholar]
- Hanson, M. A., Barreiro, P. G., Crosetto, P., & Brockington, D. (2024). The strain on scientific publishing. Quantitative Science Studies, 5(4), 823–843. [Google Scholar] [CrossRef]
- Huang, L., Yu, W., Ma, W., Zhong, W., Feng, Z., Wang, H., Chen, Q., Peng, W., Feng, X., Qin, B., & Liu, T. (2025). A survey on hallucination in large language models: Principles, taxonomy, challenges, and open questions. ACM Transactions on Information Systems, 43(2), 1–55. [Google Scholar] [CrossRef]
- Jin, S. (2024). Should we publish fewer papers? ACS Energy Letters, 9(8), 4196–4198. [Google Scholar] [CrossRef]
- King, M. R. (2023). A place for large language models in scientific publishing, apart from credited authorship. Cellular and Molecular Bioengineering, 16(2), 95–98. [Google Scholar] [CrossRef] [PubMed]
- Knoth, P., Herrmannova, D., Cancellieri, M., Anastasiou, L., Pontika, N., Pearce, S., Gyawali, B., & Pride, D. (2023). CORE: A global aggregation service for open access papers. Nature Scientific Data, 10(1), 366. [Google Scholar]
- Kreutz, C. K., & Schenkel, R. (2022). Scientific paper recommendation systems: A literature review of recent publications. arXiv, arXiv:2201.00682. Available online: https://arxiv.org/abs/2201.00682 (accessed on 20 February 2025).
- Lubos, S., Tran, T. N. T., Felfernig, A., Polat Erdeniz, S., & Le, V.-M. (2024). LLM-generated explanations for recommender systems. In Adjunct Proceedings of the 32nd ACM Conference on User Modeling, Adaptation and Personalization, Cagliari, Italy, July 1–4 (pp. 276–285). Association for Computing Machinery. [Google Scholar] [CrossRef]
- Marvin, G., Hellen, N., Jjingo, D., & Nakatumba-Nabende, J. (2023, June 27–28). Prompt engineering in large language models. International Conference on Data Intelligence and Cognitive Informatics (pp. 387–402), Tirunelveli, India. [Google Scholar]
- Nguyen, V. K. (2024). Interdisciplinary trends in current research and development of science and technology. VNU Journal of Science: Policy and Management Studies, 40(1), 1–15. [Google Scholar]
- Pranckutė, R. (2021). Web of Science (WoS) and scopus: The titans of bibliographic information in today’s academic world. Publications, 9(1), 12. [Google Scholar] [CrossRef]
- Pu, X., Gao, M., & Wan, X. (2023). Summarization is (almost) dead. arXiv, arXiv:2309.09558. Available online: https://arxiv.org/abs/2309.09558 (accessed on 20 February 2025).
- Publications, Volume 12, Issue 1. (2024). Available online: https://www.mdpi.com/2304-6775/12/1 (accessed on 20 February 2025).
- Rosman, T., Mayer, A. K., & Krampen, G. (2016). On the pitfalls of bibliographic database searching: Comparing successful and less successful users. Behaviour & Information Technology, 35(2), 106–117. [Google Scholar] [CrossRef]
- Santini, F. (2024). The scientific system must bend to avoid breaking. European Radiology, 34(9), 5866–5867. [Google Scholar] [PubMed]
- Shin, M., Soen, A., Readshaw, B. T., Blackburn, S. M., Whitelaw, M., & Xie, L. (2019, October 20–25). Influence flowers of academic entities. 2019 IEEE Conference on Visual Analytics Science and Technology (VAST) (pp. 1–10), Vancouver, BC, Canada. [Google Scholar] [CrossRef]
- Shmueli, G., Maria Colosimo, B., Martens, D., Padman, R., Saar-Tsechansky, M., R. Liu Sheng, O., Street, W. N., & Tsui, K.-L. (2023). How can IJDS authors, reviewers, and editors use (and misuse) generative AI? INFORMS Journal on Data Science, 2(1), 1–9. [Google Scholar]
- Tsai, A. Y., Kraft, A., Jin, L., Cai, C., Hosseini, A., Xu, T., Zhang, Z., Hong, L., Chi, E. H., & Yi, X. (2024). Leveraging llm reasoning enhances personalized recommender systems. Available online: https://arxiv.org/abs/2408.00802 (accessed on 20 February 2025).
- Villalobos, P., Ho, A., Sevilla, J., Besiroglu, T., Heim, L., & Hobbhahn, M. (2022). Will we run out of data? Limits of LLM scaling based on human-generated data. arXiv, arXiv:2211.04325. [Google Scholar]
- Wang, Q., Li, J., Wang, S., Xing, Q., Niu, R., Kong, H., Li, R., Long, G., Chang, Y., & Zhang, C. (2024). Towards next-generation llm-based recommender systems: A survey and beyond. Available online: https://arxiv.org/abs/2410.19744 (accessed on 20 February 2025).
- Wu, Y. (2024). Large Language Model and Text Generation. In H. Xu, & D. Demner Fushman (Eds.), Natural language processing in biomedicine: A practical guide (pp. 265–297). Springer International Publishing. [Google Scholar] [CrossRef]
- Zakkas, P., Verberne, S., & Zavrel, J. (2024). SumBlogger: Abstractive summarization of large collections of scientific articles. In N. Goharian, N. Tonellotto, Y. He, A. Lipani, G. McDonald, C. Macdonald, & I. Ounis (Eds.), Advances in information retrieval (pp. 371–386). Springer Nature Switzerland. [Google Scholar]
- Zhang, Z., Patra, B. G., Yaseen, A., Zhu, J., Sabharwal, R., Roberts, K., Cao, T., & Wu, H. (2023). Scholarly recommendation systems: A literature survey. Knowledge and Information Systems, 65, 4433–4478. [Google Scholar] [CrossRef]
- Zhao, Z., Song, S., Duah, B., Macbeth, J., Carter, S., Van, M. P., Bravo, N. S., Klenk, M., Sick, K., & Filipowicz, A. L. (2023, June 19–21). More human than human: LLM-generated narratives outperform human-LLM interleaved narratives. 15th Conference on Creativity and Cognition (pp. 368–370), Virtual. [Google Scholar]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the author. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).