Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

Article Types

Countries / Regions

Search Results (79)

Search Parameters:
Keywords = human-chatbot interaction

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
59 pages, 4837 KB  
Article
A Human–AI Compass for Sustainable Art Museums: Navigating Opportunities and Challenges in Operations, Collections Management, and Visitor Engagement
by Charis Avlonitou, Eirini Papadaki and Alexandros Apostolakis
Heritage 2025, 8(10), 422; https://doi.org/10.3390/heritage8100422 - 5 Oct 2025
Viewed by 731
Abstract
This paper charts AI’s transformative path toward advancing sustainability within art museums, introducing a Human–AI compass as a conceptual framework for navigating its integration. It advocates for human-centric AI that optimizes operations, modernizes collection management, and deepens visitor engagement—anchored in meaningful human–technology synergy [...] Read more.
This paper charts AI’s transformative path toward advancing sustainability within art museums, introducing a Human–AI compass as a conceptual framework for navigating its integration. It advocates for human-centric AI that optimizes operations, modernizes collection management, and deepens visitor engagement—anchored in meaningful human–technology synergy and thoughtful human oversight. Drawing on extensive literature review and real-world museum case studies, the paper explores AI’s multifaceted impact across three domains. Firstly, it examines how AI improves operations, from audience forecasting and resource optimization to refining marketing, supporting conservation, and reshaping curatorial practices. Secondly, it investigates AI’s influence on digital collection management, highlighting its ability to improve organization, searchability, analysis, and interpretation through automated metadata and advanced pattern recognition. Thirdly, the study analyzes how AI elevates the visitor experience via chatbots, audio guides, and interactive applications, leveraging personalization, recommendation systems, and co-creation opportunities. Crucially, this exploration acknowledges AI’s complex challenges—technical-operational, ethical-governance, socioeconomic-cultural, and environmental—underscoring the indispensable role of human judgment in steering its implementation. The Human-AI compass offers a balanced, strategic approach for aligning innovation with human values, ethical principles, museum mission, and sustainability. The study provides valuable insights for researchers, practitioners and policymakers, enriching the broader discourse on AI’s growing role in the art and cultural sector. Full article
Show Figures

Figure 1

42 pages, 1748 KB  
Article
Memory-Augmented Large Language Model for Enhanced Chatbot Services in University Learning Management Systems
by Jaeseung Lee and Jehyeok Rew
Appl. Sci. 2025, 15(17), 9775; https://doi.org/10.3390/app15179775 - 5 Sep 2025
Viewed by 1473
Abstract
A learning management system (LMS) plays a crucial role in supporting students’ educational activities by centralized platforms for course delivery, communication, and student support. Recently, many universities have integrated chatbots into their LMS to assist students with various inquiries and tasks. However, existing [...] Read more.
A learning management system (LMS) plays a crucial role in supporting students’ educational activities by centralized platforms for course delivery, communication, and student support. Recently, many universities have integrated chatbots into their LMS to assist students with various inquiries and tasks. However, existing chatbots often necessitate human interventions to manually respond to complex queries, resulting in limited scalability and efficiency. In this paper, we present a memory-augmented large language model (LLM) framework that enhances the reasoning and contextual continuity of LMS-based chatbots. The proposed framework first embeds user queries and retrieves semantically relevant entries from various LMS resources, including instructional documents and academic frequently asked questions. Retrieved entries are then filtered through a two-stage confidence filtering process that combines similarity thresholds and LLM-based semantic validation. Validated information, along with user queries, is processed by LLM for response generation. To maintain coherence in multi-turn interactions, the chatbot incorporates short-term, long-term, and temporal event memories, which track conversational flow and personalize responses based on user-specific information, such as recent activity history and individual preferences. To evaluate response quality, we employed a multi-layered evaluation strategy combining BERTScore-based quantitative measurement, an LLM-as-a-Judge approach for automated semantic assessment, and a user study under multi-turn scenarios. The evaluation results consistently confirm that the proposed framework improves the consistency, clarity, and usefulness of the responses. These findings highlight the potential of memory-augmented LLMs for scalable and intelligent learning support within university environments. Full article
(This article belongs to the Special Issue Applications of Digital Technology and AI in Educational Settings)
Show Figures

Figure 1

22 pages, 693 KB  
Article
How Perceived Motivations Influence User Stickiness and Sustainable Engagement with AI-Powered Chatbots—Unveiling the Pivotal Function of User Attitude
by Hua Pang, Zhuyun Hu and Lei Wang
J. Theor. Appl. Electron. Commer. Res. 2025, 20(3), 228; https://doi.org/10.3390/jtaer20030228 - 1 Sep 2025
Viewed by 806
Abstract
Artificial intelligence (AI) is reshaping customer service, with AI-powered chatbots serving as a critical component in delivering continuous support across sales, marketing, and service domains, thereby enhancing operational efficiency. However, consumer engagement remains suboptimal, as many users favor human interaction due to concerns [...] Read more.
Artificial intelligence (AI) is reshaping customer service, with AI-powered chatbots serving as a critical component in delivering continuous support across sales, marketing, and service domains, thereby enhancing operational efficiency. However, consumer engagement remains suboptimal, as many users favor human interaction due to concerns regarding chatbots’ ability to address complex issues and their perceived lack of empathy, which subsequently reduces satisfaction and sustainable usage. This study examines the determinants of user attitude and identifies factors influencing sustainable chatbot use. Utilizing survey data from 735 Chinese university students who have engaged with AI-powered chatbots, the analysis reveals that four key motivational categories: utilitarian (information acquisition), hedonic (enjoyment and time passing), technology (media appeal), and social (social presence and interaction) significantly influence user attitude toward chatbot services. Conversely, privacy invasion exerts a negative impact on user attitude, suggesting that while chatbots provide certain benefits, privacy issues can significantly undermine user satisfaction. Moreover, the findings suggest that user attitude serves as a pivotal determinant in fostering both user stickiness and sustainable usage of chatbot services. This study advances prior U&G-, TAM-, and ECM-based research by applying these frameworks to AI-powered chatbots in business communication, refining the U&G model with four specific motivations, integrating perceived privacy invasion to bridge gratification theory with risk perception, and directly linking user motivations to business outcomes such as attitude and stickiness. This study underscores that optimizing chatbot functionalities to enhance user gratification while mitigating privacy risks can substantially improve user satisfaction and stickiness, offering valuable implications for businesses aiming to enhance customer loyalty through AI-powered services. Full article
Show Figures

Figure 1

23 pages, 994 KB  
Article
Driving Consumer Engagement Through AI Chatbot Experience: The Mediating Role of Satisfaction Across Generational Cohorts and Gender in Travel Tourism
by José Magano, Joana A. Quintela and Neelotpaul Banerjee
Sustainability 2025, 17(17), 7673; https://doi.org/10.3390/su17177673 - 26 Aug 2025
Viewed by 1790
Abstract
This study explores how AI chatbot experiences on travel websites influence consumer engagement, with satisfaction from using AI chatbots as a mediating factor. Grounded in the Stimulus-Organism-Response (S-O-R) framework, the research shifts the focus from utilitarian models to examine how chatbot attributes—e.g., ease [...] Read more.
This study explores how AI chatbot experiences on travel websites influence consumer engagement, with satisfaction from using AI chatbots as a mediating factor. Grounded in the Stimulus-Organism-Response (S-O-R) framework, the research shifts the focus from utilitarian models to examine how chatbot attributes—e.g., ease of use, information quality, security, anthropomorphism, and omnipresence—affect satisfaction of using AI chatbots and subsequent consumer engagement behaviours. Survey data from 519 Portuguese travellers were analysed using partial least squares structural equation modelling (PLS-SEM). The study contributes to theory by (1) demonstrating S-O-R’s advantages over utilitarian models in capturing relational and emotional dimensions of AI interactions, (2) identifying satisfaction with using AI chatbots as a pivotal mediator between AI chatbot experience and consumer engagement, and (3) revealing generational disparities in drivers of engagement. Notably, satisfaction strongly influences engagement for Generation X, while direct experience matters more for Generation Z. Millennials exhibit a distinct preference for hybrid human–AI service handoffs. The practical implications include prioritizing natural language processing for ease of use, implementing generational customization (e.g., gamification for Gen Z, reliability assurances for Gen X), and ensuring seamless human escalation for Millennials. These insights equip travel businesses to design AI chatbots that foster long-term loyalty and competitive differentiation. Full article
Show Figures

Figure 1

25 pages, 953 KB  
Article
Communication Errors in Human–Chatbot Interactions: A Case Study of ChatGPT Arabic Mental Health Support Inquiries
by Ghuzayyil Mohammed Al-Otaibi, Hind M. Alotaibi and Sami Sulaiman Alsalmi
Behav. Sci. 2025, 15(8), 1119; https://doi.org/10.3390/bs15081119 - 18 Aug 2025
Viewed by 867
Abstract
Large language models (LLMs) have become extensively used among users across diverse settings. Yet, with the complex nature of these large-scale artificial intelligence (AI) systems, leveraging their capabilities effectively is yet to be explored. In this study, we looked at the types of [...] Read more.
Large language models (LLMs) have become extensively used among users across diverse settings. Yet, with the complex nature of these large-scale artificial intelligence (AI) systems, leveraging their capabilities effectively is yet to be explored. In this study, we looked at the types of communication errors that occur in interactions between humans and ChatGPT-3.5 in Arabic. A corpus of six Arabic-language consultations was collected from an online mental health support forum. For each consultation, the researchers provided the user’s Arabic queries to ChatGPT-3.5 and analyzed the system’s responses. The study identified 102 communication errors, mostly grammatical and repetitions. Other errors involved contradictions, ambiguous language, ignoring questions, and lacking sociality. By examining the patterns and types of communication errors observed in ChatGPT’s responses, the study is expected to provide insights into the challenges and limitations of current conversational AI systems, particularly in the context of sensitive domains like mental health support. Full article
(This article belongs to the Special Issue Digital Interventions for Addiction and Mental Health)
Show Figures

Figure 1

22 pages, 1780 KB  
Systematic Review
The Future of Education: A Systematic Literature Review of Self-Directed Learning with AI
by Carmen del Rosario Navas Bonilla, Luis Miguel Viñan Carrasco, Jhoanna Carolina Gaibor Pupiales and Daniel Eduardo Murillo Noriega
Future Internet 2025, 17(8), 366; https://doi.org/10.3390/fi17080366 - 13 Aug 2025
Cited by 2 | Viewed by 2977
Abstract
As digital transformation continues to redefine education, understanding how emerging technologies can enhance self-directed learning (SDL) becomes essential for learners, educators, instructional designers, and policymakers, as this approach supports personalized learning, strengthens student autonomy, and responds to the demands of more flexible and [...] Read more.
As digital transformation continues to redefine education, understanding how emerging technologies can enhance self-directed learning (SDL) becomes essential for learners, educators, instructional designers, and policymakers, as this approach supports personalized learning, strengthens student autonomy, and responds to the demands of more flexible and dynamic educational environments. This systematic review examines how artificial intelligence (AI) tools enhance SDL by offering personalized, adaptive, and real-time support for learners in online environments. Following the PRISMA 2020 methodology, a literature search was conducted to identify relevant studies published between 2020 and 2025. After applying inclusion, exclusion, and quality criteria, 77 studies were selected for in-depth analysis. The findings indicate that AI-powered tools such as intelligent tutoring systems, chatbots, conversational agents, and natural language processing applications promote learner autonomy, enable self-regulation, provide real-time feedback, and support individualized learning paths. However, several challenges persist, including overreliance on technology, cognitive overload, and diminished human interaction. These insights suggest that, while AI plays a transformative role in the evolution of education, its integration must be guided by thoughtful pedagogical design, ethical considerations, and a learner-centered approach to fully support the future of education through the internet. Full article
Show Figures

Figure 1

28 pages, 1622 KB  
Article
Trusting Humans or Bots? Examining Trust Transfer and Algorithm Aversion in China’s E-Government Services
by Yifan Song, Takashi Natori and Xintao Yu
Adm. Sci. 2025, 15(8), 308; https://doi.org/10.3390/admsci15080308 - 6 Aug 2025
Cited by 1 | Viewed by 1756
Abstract
Despite the increasing integration of government chatbots (GCs) into digital public service delivery, their real-world effectiveness remains limited. Drawing on the literature on algorithm aversion, trust-transfer theory, and perceived risk theory, this study investigates how the type of service agent (human vs. GCs) [...] Read more.
Despite the increasing integration of government chatbots (GCs) into digital public service delivery, their real-world effectiveness remains limited. Drawing on the literature on algorithm aversion, trust-transfer theory, and perceived risk theory, this study investigates how the type of service agent (human vs. GCs) influences citizens’ trust of e-government services (TOE) and e-government service adoption intention (EGA). Furthermore, it explores whether the effect of trust of government (TOG) on TOE differs across agent types, and whether perceived risk (PR) serves as a boundary condition in this trust-transfer process. An online scenario-based experiment was conducted with a sample of 318 Chinese citizens. Data were analyzed using the Mann–Whitney U test and partial least squares structural equation modeling (PLS-SEM). The results reveal that, within the Chinese e-government context, citizens perceive higher risk (PR) and report lower adoption intention (EGA) when interacting with GCs compared to human agents—an indication of algorithm aversion. However, high levels of TOG mitigate this aversion by enhancing TOE. Importantly, PR moderates the strength of this trust-transfer effect, serving as a critical boundary condition. Full article
(This article belongs to the Special Issue Challenges and Future Trends in Digital Government)
Show Figures

Figure 1

14 pages, 283 KB  
Article
Teens, Tech, and Talk: Adolescents’ Use of and Emotional Reactions to Snapchat’s My AI Chatbot
by Gaëlle Vanhoffelen, Laura Vandenbosch and Lara Schreurs
Behav. Sci. 2025, 15(8), 1037; https://doi.org/10.3390/bs15081037 - 30 Jul 2025
Viewed by 2932
Abstract
Due to technological advancements such as generative artificial intelligence (AI) and large language models, chatbots enable increasingly human-like, real-time conversations through text (e.g., OpenAI’s ChatGPT) and voice (e.g., Amazon’s Alexa). One AI chatbot that is specifically designed to meet the social-supportive needs of [...] Read more.
Due to technological advancements such as generative artificial intelligence (AI) and large language models, chatbots enable increasingly human-like, real-time conversations through text (e.g., OpenAI’s ChatGPT) and voice (e.g., Amazon’s Alexa). One AI chatbot that is specifically designed to meet the social-supportive needs of youth is Snapchat’s My AI. Given its increasing popularity among adolescents, the present study investigated whether adolescents’ likelihood of using My AI, as well as their positive or negative emotional experiences from interacting with the chatbot, is related to socio-demographic factors (i.e., gender, age, and socioeconomic status (SES)). A cross-sectional study was conducted among 303 adolescents (64.1% girls, 35.9% boys, 1.0% other, 0.7% preferred not to say their gender; Mage = 15.89, SDage = 1.69). The findings revealed that younger adolescents were more likely to use My AI and experienced more positive emotions from these interactions than older adolescents. No significant relationships were found for gender or SES. These results highlight the potential for age to play a critical role in shaping adolescents’ engagement with AI chatbots on social media and their emotional outcomes from such interactions, underscoring the need to consider developmental factors in AI design and policy. Full article
40 pages, 759 KB  
Systematic Review
Decoding Trust in Artificial Intelligence: A Systematic Review of Quantitative Measures and Related Variables
by Letizia Aquilino, Cinzia Di Dio, Federico Manzi, Davide Massaro, Piercosma Bisconti and Antonella Marchetti
Informatics 2025, 12(3), 70; https://doi.org/10.3390/informatics12030070 - 14 Jul 2025
Viewed by 4521
Abstract
As artificial intelligence (AI) becomes ubiquitous across various fields, understanding people’s acceptance and trust in AI systems becomes essential. This review aims to identify quantitative measures used to measure trust in AI and the associated studied elements. Following the PRISMA guidelines, three databases [...] Read more.
As artificial intelligence (AI) becomes ubiquitous across various fields, understanding people’s acceptance and trust in AI systems becomes essential. This review aims to identify quantitative measures used to measure trust in AI and the associated studied elements. Following the PRISMA guidelines, three databases were consulted, selecting articles published before December 2023. Ultimately, 45 articles out of 1283 were selected. Articles were included if they were peer-reviewed journal publications in English reporting empirical studies measuring trust in AI systems with multi-item questionnaires. Studies were analyzed through the lenses of cognitive and affective trust. We investigated trust definitions, questionnaires employed, types of AI systems, and trust-related constructs. Results reveal diverse trust conceptualizations and measurements. In addition, the studies covered a wide range of AI system types, including virtual assistants, content detection tools, chatbots, medical AI, robots, and educational AI. Overall, the studies show compatibility of cognitive or affective trust focus between theorization, items, experimental stimuli, and level of anthropomorphism of the systems. The review underlines the need to adapt measurement of trust in the specific characteristics of human–AI interaction, accounting for both the cognitive and affective sides. Trust definitions and measurement could be chosen depending also on the level of anthropomorphism of the systems and the context of application. Full article
Show Figures

Figure 1

17 pages, 299 KB  
Article
Analysis of Reliability and Efficiency of Information Extraction Using AI-Based Chatbot: The More-for-Less Paradox
by Eugene Levner and Boris Kriheli
Algorithms 2025, 18(7), 412; https://doi.org/10.3390/a18070412 - 3 Jul 2025
Viewed by 535
Abstract
This paper addresses the problem of information extraction using an AI-powered chatbot. The problem concerns searching and extracting relevant information from large databases in response to a human user’s query. Expanding the traditional discrete search problem well known in operations research, this problem [...] Read more.
This paper addresses the problem of information extraction using an AI-powered chatbot. The problem concerns searching and extracting relevant information from large databases in response to a human user’s query. Expanding the traditional discrete search problem well known in operations research, this problem introduces two players; the first player—an AI chatbot such as ChatGPT 4.0—sequentially scans available datasets to find an appropriate answer to a given query, while the second—a human user—conducts a dialogue with the chatbot and evaluates its answers in each round of the dialogue. The goal of an AI-powered chatbot is to provide maximally useful and accurate information. During a natural language conversation between a human user and an AI, the human user can modify and refine queries until s/he is satisfied with the chatbot’s output. We analyze two key characteristics of human–AI interaction: search reliability and efficiency. Search reliability is defined as the ability of a robot to understand user queries and provide correct answers; it is measured by the frequency (probability) of correct answers. Search efficiency of a chatbot indicates how accurate and relevant the information returned by the chatbot is; it is measured by the satisfaction level a human user receives for a correct answer. An AI chatbot must perform a sequence of scans over the given databases and continue searching until the human user declares, in some round, that the target has been found. Assuming that the chatbot is not completely reliable, each database may have to be scanned infinitely often; in this case, the objective of the problem is to determine a search policy for finding the optimal sequence of chatbot scans that maximizes the expected user satisfaction over an infinite time horizon. Along with these results, we found a counterintuitive relationship between AI chatbot reliability and search performance: under sufficiently general conditions, a less reliable AI chatbot may have higher expected search efficiency; this phenomenon aligns with other well-known “more-for-less” paradoxes. Finally, we discussed the underlying mechanism of this paradox. Full article
24 pages, 1501 KB  
Review
Large Language Models in Medical Chatbots: Opportunities, Challenges, and the Need to Address AI Risks
by James C. L. Chow and Kay Li
Information 2025, 16(7), 549; https://doi.org/10.3390/info16070549 - 27 Jun 2025
Cited by 4 | Viewed by 4555
Abstract
Large language models (LLMs) are transforming the capabilities of medical chatbots by enabling more context-aware, human-like interactions. This review presents a comprehensive analysis of their applications, technical foundations, benefits, challenges, and future directions in healthcare. LLMs are increasingly used in patient-facing roles, such [...] Read more.
Large language models (LLMs) are transforming the capabilities of medical chatbots by enabling more context-aware, human-like interactions. This review presents a comprehensive analysis of their applications, technical foundations, benefits, challenges, and future directions in healthcare. LLMs are increasingly used in patient-facing roles, such as symptom checking, health information delivery, and mental health support, as well as in clinician-facing applications, including documentation, decision support, and education. However, as a study from 2024 warns, there is a need to manage “extreme AI risks amid rapid progress”. We examine transformer-based architectures, fine-tuning strategies, and evaluation benchmarks specific to medical domains to identify their potential to transfer and mitigate AI risks when using LLMs in medical chatbots. While LLMs offer advantages in scalability, personalization, and 24/7 accessibility, their deployment in healthcare also raises critical concerns. These include hallucinations (the generation of factually incorrect or misleading content by an AI model), algorithmic biases, privacy risks, and a lack of regulatory clarity. Ethical and legal challenges, such as accountability, explainability, and liability, remain unresolved. Importantly, this review integrates broader insights on AI safety, drawing attention to the systemic risks associated with rapid LLM deployment. As highlighted in recent policy research, including work on managing extreme AI risks, there is an urgent need for governance frameworks that extend beyond technical reliability to include societal oversight and long-term alignment. We advocate for responsible innovation and sustained collaboration among clinicians, developers, ethicists, and regulators to ensure that LLM-powered medical chatbots are deployed safely, equitably, and transparently within healthcare systems. Full article
Show Figures

Graphical abstract

15 pages, 1255 KB  
Article
Do Chatbots Exhibit Personality Traits? A Comparison of ChatGPT and Gemini Through Self-Assessment
by W. Wiktor Jedrzejczak and Joanna Kobosko
Information 2025, 16(7), 523; https://doi.org/10.3390/info16070523 - 23 Jun 2025
Cited by 1 | Viewed by 2336
Abstract
The underlying design of large language models (LLMs), trained on vast amounts of human texts, implies that chatbots based on them will almost inevitably retain some human personality traits. That is, we expect that LLM outputs will tend to reflect human-like features. In [...] Read more.
The underlying design of large language models (LLMs), trained on vast amounts of human texts, implies that chatbots based on them will almost inevitably retain some human personality traits. That is, we expect that LLM outputs will tend to reflect human-like features. In this study, we used the ‘Big Five’ personality traits tool to examine whether several chatbot models (ChatGPT versions 3.5 and 4o, Gemini, and Gemini Advanced, all tested in both English and Polish), displayed distinctive personality profiles. Each chatbot was presented with an instruction to complete the International Personality Item Pool (IPIP) questionnaire “according to who or what you are,” which left it open as to whether the answer would derive from a purported human or from an AI source. We found that chatbots sometimes chose to respond in a typically human-like way, while in other cases the answers appeared to reflect the perspective of an AI language model. The distinction was examined more closely through a set of follow-up questions. The more advanced models (ChatGPT-4o and Gemini Advanced) showed larger differences between these two modes compared to the more basic models. In IPIP-5 terms, the chatbots tended to display higher ‘Emotional Stability’ and ‘Intellect/Imagination’ but lower ‘Agreeableness’ compared to published human norms. The spread of characteristics indicates that the personality profiles of chatbots are not static but are shaped by the model architecture and its programming as well as, perhaps, the chatbot’s own inner sense, that is, the way it models its own identity. Appreciating these philosophical subtleties is important for enhancing human–computer interactions and perhaps building more relatable, trustworthy AI systems. Full article
Show Figures

Graphical abstract

21 pages, 2985 KB  
Article
Non-Invasive Fatigue Detection and Human–Machine Interaction Using LSTM and Multimodal AI: A Case Study
by Muon Ha, Yulia Shichkina and Xuan-Hien Nguyen
Multimodal Technol. Interact. 2025, 9(6), 63; https://doi.org/10.3390/mti9060063 - 13 Jun 2025
Viewed by 1413
Abstract
Fatigue in high-stress work environments poses significant risks to employee performance and safety. This study introduces a non-invasive fatigue detection system utilizing facial parameters processed via a Long Short-Term Memory (LSTM) neural network, coupled with a human–machine interaction interface via a Telegram chatbot. [...] Read more.
Fatigue in high-stress work environments poses significant risks to employee performance and safety. This study introduces a non-invasive fatigue detection system utilizing facial parameters processed via a Long Short-Term Memory (LSTM) neural network, coupled with a human–machine interaction interface via a Telegram chatbot. The system analyzes eye blink patterns and facial expression changes captured through a webcam, achieving an accuracy of 92.35% on the UTA-RLDD dataset. An interactive feedback mechanism allows users to verify predictions, enhancing system adaptability. We further propose a multimodal AI framework to integrate physiological and environmental data, laying the groundwork for broader applications. This approach provides an effective solution for early fatigue detection and adaptive collaboration between humans and machines in real-time settings. Full article
Show Figures

Figure 1

20 pages, 2898 KB  
Article
Deploying a Mental Health Chatbot in Higher Education: The Development and Evaluation of Luna, an AI-Based Mental Health Support System
by Phillip Olla, Ashlee Barnes, Lauren Elliott, Mustafa Abumeeiz, Venus Olla and Joseph Tan
Computers 2025, 14(6), 227; https://doi.org/10.3390/computers14060227 - 10 Jun 2025
Viewed by 2976
Abstract
Rising mental health challenges among postsecondary students have increased the demand for scalable, ethical solutions. This paper presents the design, development, and safety evaluation of Luna, a GPT-4-based mental health chatbot. Built using a modular PHP architecture, Luna integrates multi-layered prompt engineering, safety [...] Read more.
Rising mental health challenges among postsecondary students have increased the demand for scalable, ethical solutions. This paper presents the design, development, and safety evaluation of Luna, a GPT-4-based mental health chatbot. Built using a modular PHP architecture, Luna integrates multi-layered prompt engineering, safety guardrails, and referral logic. The Institutional Review Board (IRB) at the University of Detroit Mercy (Protocol #23-24-38) reviewed the proposed study and deferred full human subject approval, requesting technical validation prior to deployment. In response, we conducted a pilot test with a variety of users—including clinicians and students who simulated at-risk student scenarios. Results indicated that 96% of expert interactions were deemed safe, and 90.4% of prompts were considered useful. This paper describes Luna’s architecture, prompt strategy, and expert feedback, concluding with recommendations for future human research trials. Full article
Show Figures

Figure 1

17 pages, 1910 KB  
Article
AI Response Quality in Public Services: Temperature Settings and Contextual Factors
by Domenico Trezza, Giuseppe Luca De Luca Picione and Carmine Sergianni
Societies 2025, 15(5), 127; https://doi.org/10.3390/soc15050127 - 6 May 2025
Viewed by 1026
Abstract
This study investigated how generative Artificial Intelligence (AI) systems—now increasingly integrated into public services—respond to different technical configurations, and how these configurations affect the perceived quality of the outputs. Drawing on an experimental evaluation of Govern-AI, a chatbot designed for professionals in [...] Read more.
This study investigated how generative Artificial Intelligence (AI) systems—now increasingly integrated into public services—respond to different technical configurations, and how these configurations affect the perceived quality of the outputs. Drawing on an experimental evaluation of Govern-AI, a chatbot designed for professionals in the social, educational, and labor sectors, we analyzed the impact of the temperature parameter—which controls the degree of creativity and variability in the responses—on two key dimensions: accuracy and comprehensibility. This analysis was based on 8880 individual evaluations collected from five professional profiles. The findings revealed the following: (1) the high-temperature responses were generally more comprehensible and appreciated, yet less accurate in strategically sensitive contexts; (2) professional groups differed significantly in their assessments, where trade union representatives and regional policy staff expressed more critical views than the others; (3) the type of question—whether operational or informational—significantly influenced the perceived output quality. This study demonstrated that the AI performance was far from neutral: it depended on technical settings, usage contexts, and the profiles of the end users. Investigating these “behind-the-scenes” dynamics is essential for fostering the informed governance of AI in public services, and for avoiding the risk of technology functioning as an opaque black box within decision-making processes. Full article
Show Figures

Figure 1

Back to TopTop