Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (1,315)

Search Parameters:
Keywords = ChatGPT

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
27 pages, 1145 KB  
Article
Something Old, Something New: WebQuests and GenAI in Teacher Education
by Peter Tiernan, Enda Donlon, Mahmoud Hamash and James Lovatt
AI Educ. 2026, 2(1), 7; https://doi.org/10.3390/aieduc2010007 - 11 Mar 2026
Abstract
Generative artificial intelligence (GenAI) has rapidly emerged as a transformative educational technology, raising questions about how educators and pre-service teachers critically engage with AI-produced content. This case study investigates how WebQuests, a long-established, inquiry-based pedagogical model, can foster critical engagement with GenAI tools. [...] Read more.
Generative artificial intelligence (GenAI) has rapidly emerged as a transformative educational technology, raising questions about how educators and pre-service teachers critically engage with AI-produced content. This case study investigates how WebQuests, a long-established, inquiry-based pedagogical model, can foster critical engagement with GenAI tools. Situated within an initial teacher education programme, a WebQuest, incorporating GenAI sources, was implemented with 24 pre-service language teachers, who engaged with curated resources alongside ChatGPT and Copilot to produce infographics for secondary school audiences. Data were collected through semi-structured interviews and were analysed using Braun and Clarke’s thematic analysis. Findings indicate that scaffolded engagement with GenAI encouraged participants to compare AI-generated outputs with trusted sources, critically evaluate accuracy and reliability, and reflect on integration into their future practice. Whilst pre-service teachers valued GenAI’s accessibility and efficiency, they expressed concerns about clarity, verbosity, and trustworthiness. The WebQuest model effectively supported synthesis of multiple information sources, fostering functional AI engagement and critical evaluation of its affordances and limitations. This case study concludes that integrating GenAI within structured, inquiry-based pedagogies advances digital and AI literacy in initial teacher education, whilst highlighting the need for institutional guidance, professional development, and further research in this area. Full article
Show Figures

Figure 1

27 pages, 1113 KB  
Article
On the Investigation of Environmental Effects of ChatGPT Usage via the Newly Developed Mathematical Model in Caputo Sense
by Sherly K, Pundikala Veeresha and Haci Mehmet Baskonus
Fractal Fract. 2026, 10(3), 184; https://doi.org/10.3390/fractalfract10030184 - 11 Mar 2026
Abstract
This study explores the interconnection between the variables of ChatGPT usage, energy consumption, water consumption, and carbon dioxide CO2 emissions. A new integer and fractional order model using the Caputo derivative is proposed to comprehend the long-term dependencies of these variables. Boundedness, [...] Read more.
This study explores the interconnection between the variables of ChatGPT usage, energy consumption, water consumption, and carbon dioxide CO2 emissions. A new integer and fractional order model using the Caputo derivative is proposed to comprehend the long-term dependencies of these variables. Boundedness, and global and local stability are examined for the fractional order model. The equilibrium points of these variables are shown to determine the stability of the model. The Runge–Kutta 7 numerical method is employed for the integer order model, whereas the semi-implicit linear interpolation (L1) method is used for the fractional order model. The parameter sensitivity is conducted on the system’s parameters to understand the variables’ impact by varying the relevant parameters for the system. To increase the efficacy of our analysis, we used machine learning approaches to model and predict the dynamics of CO2 emissions, energy and water consumption, and ChatGPT usage. The Prophet ML model stood out among the other methods because it is adept at identifying long-term growth trends, seasonal changes, and the impact of outside variables in intricate time-series data. It is extremely beneficial for research centered on sustainability, where accurate projections are essential for wellinformed decision-making, because it can produce robust, interpretable forecasts against missing values and outliers. Using the Prophet ML model, our research guarantees precise and expandable predictions and provides valuable information that can direct tactics to balance environmental sustainability and technological progress. Full article
Show Figures

Figure 1

13 pages, 233 KB  
Article
Quality and Usability of Prostate Cancer Information Generated by Artificial Intelligence Chatbots: A Comparative Analysis
by Abdullah Al-Khanaty, Jordan Santucci, David Hennes, Niranjan Sathianathen, Carlos Delgado, Karan Sharma, Eoin Dinneen, Kieran Sandhu, David Chen, Renu Eapen, Daniel Moon, Gregory Jack, Jeremy Goad, Shankar Siva, Muhammad Ali, Damien Bolton, Nathan Lawrentschuk, Declan G. Murphy and Marlon Perera
Cancers 2026, 18(6), 906; https://doi.org/10.3390/cancers18060906 - 11 Mar 2026
Abstract
Background: Artificial intelligence chatbots are increasingly used by patients to obtain health information, including for prostate cancer. While these platforms offer accessible and conversational responses, concerns remain regarding the quality, usability, and clinical relevance of AI-generated content. This study comparatively evaluated patient-directed prostate [...] Read more.
Background: Artificial intelligence chatbots are increasingly used by patients to obtain health information, including for prostate cancer. While these platforms offer accessible and conversational responses, concerns remain regarding the quality, usability, and clinical relevance of AI-generated content. This study comparatively evaluated patient-directed prostate cancer information generated by commonly used AI chatbots. Methods: Standardised prostate cancer-related prompts were developed using Google Trends and authoritative healthcare resources. Identical queries were submitted to five publicly accessible AI chatbots: ChatGPT 5.2, Google Gemini, Claude AI, Microsoft Copilot, and Perplexity. Responses were independently assessed by two blinded reviewers using the DISCERN instrument for information quality and the Patient Education Materials Assessment Tool for printable materials (PEMAT-P) for understandability and actionability. Inter-rater reliability was assessed using intraclass correlation coefficients (ICCs). Readability was evaluated using the Flesch–Kincaid Reading Ease score. Descriptive statistics were used for comparative and pooled analyses. Results: Overall information quality was moderate, with a pooled median (interquartile range [IQR]) DISCERN score of 56.5 (53.0–61.0). Higher mean DISCERN scores were observed for ChatGPT 5.2 and Microsoft Copilot, whereas lower scores were observed for Claude and Perplexity. PEMAT-P understandability was consistently high across platforms, with a pooled median (IQR) score of 91.7% (83.3–91.7%). In contrast, PEMAT-P actionability was uniformly poor, with a pooled median (IQR) score of 0% (0–0%). Readability analysis demonstrated moderate complexity, with a pooled median (IQR) Flesch–Kincaid Reading Ease score of 50.4 (49.2–52.5) and a median word count of 666 (657–1022). Inter-rater reliability was good for PEMAT understandability (ICC 0.841) and moderate for DISCERN (ICC 0.712). Conclusions: AI chatbots provide highly understandable but only moderately high-quality patient-directed prostate cancer information, with a consistent lack of actionable guidance. Although variation in content quality was observed across platforms, significant limitations remain in evidence transparency and practical patient support. Future development should prioritise integration of evidence-based resources and actionable decision-support tools to enhance the role of AI chatbots in prostate cancer education. Full article
11 pages, 708 KB  
Article
Evaluation of Artificial Intelligence as a Decision-Support Tool in Urological Tumor Boards: A Study in Real Clinical Practice
by Javier De la Torre-Trillo, Yaiza Yáñez Castillo, Maria Teresa Melgarejo Segura, Elisa Carmona Sánchez, Alberto Zambudio Munuera, Juan Mora-Delgado and Alfonso López Luque
J. Clin. Med. 2026, 15(6), 2130; https://doi.org/10.3390/jcm15062130 - 11 Mar 2026
Abstract
Background/Objectives: Artificial intelligence (AI) tools, particularly large language models (LLMs) such as ChatGPT-4o, are gaining prominence in medicine. While their diagnostic capabilities have been explored across various oncologic domains, their role in clinical decision-making within multidisciplinary tumor boards (MTBs) remains largely unexamined [...] Read more.
Background/Objectives: Artificial intelligence (AI) tools, particularly large language models (LLMs) such as ChatGPT-4o, are gaining prominence in medicine. While their diagnostic capabilities have been explored across various oncologic domains, their role in clinical decision-making within multidisciplinary tumor boards (MTBs) remains largely unexamined in urologic oncology. This study evaluates the performance of ChatGPT-4o as a decision-support tool in a real-world MTB setting by comparing its recommendations with those of expert clinicians. Materials and Methods: A retrospective study was conducted using 98 anonymized clinical cases discussed by a urologic MTB between June 2024 and February 2025. An independent urologist entered the same cases into ChatGPT-4o using a standardized prompt replicating real-world presentation. Two certified urologists independently assessed the model’s responses. Agreement was analyzed overall and by tumor type, disease stage, clinical context, and treatment strategy. Results: ChatGPT-4o fully agreed with the MTB in 56.1% of cases, was correct but incomplete in 23.5%, and provided partially accurate but flawed recommendations in 18.4%. Overall concordance between ChatGPT-4o and the MTB yielded a Cohen’s kappa of 0.61, indicating moderate-to-good agreement. Discrepancies were most common in metastatic prostate cancer, often due to misclassification of tumor burden or errors in treatment sequencing. Highest agreement rates were observed in bladder and renal tumors, and in standardized therapeutic scenarios such as radiotherapy. Conclusions: ChatGPT-4o demonstrated moderate alignment with expert MTB decisions and performed best in well-defined clinical contexts. While it cannot replace multidisciplinary expertise, it may serve as a supportive tool to enhance access to standardized oncologic care. Full article
Show Figures

Figure 1

7 pages, 378 KB  
Proceeding Paper
Optimizing Document Interaction Using Large Language Models by Integrating Retrieval-Augmented Generation, Facebook AI Similarity Search, and Human-like Performance Metrics
by Edwina Hon Kai Xin, Zhi Wei Tan, Ling Hue Wee and Chi Wee Tan
Eng. Proc. 2026, 128(1), 18; https://doi.org/10.3390/engproc2026128018 - 10 Mar 2026
Abstract
We developed an intelligent conversational system that enhances document interaction using advanced language models and embedding techniques. The system integrates retrieval-augmented generation, Facebook AI similarity search-based retrieval, and cosine similarity for efficient information extraction from Portable Document Format documents. It employs three embedding [...] Read more.
We developed an intelligent conversational system that enhances document interaction using advanced language models and embedding techniques. The system integrates retrieval-augmented generation, Facebook AI similarity search-based retrieval, and cosine similarity for efficient information extraction from Portable Document Format documents. It employs three embedding models, namely All-MiniLM L6 v2, All-MPNet Base v2, and Instructor Large, with three large language models including LLaMA 3.3 70B, Gemma 2-9B IT, and Mixtral 8x7B-32768. System performance is evaluated using ROUGE-1, BERTScore, and a novel human-like performance (HLP) metric, showing improved retrieval accuracy, response coherence, and efficiency for academic and enterprise applications. Full article
Show Figures

Figure 1

38 pages, 2312 KB  
Article
Transforming Learning: Use of the 4PADAFE Instructional Design Methodology and Generative Artificial Intelligence in Designing MOOCs for Innovative Education
by Lena Ivannova Ruiz-Rojas and Patricia Acosta-Vargas
Sustainability 2026, 18(6), 2683; https://doi.org/10.3390/su18062683 - 10 Mar 2026
Abstract
This study investigates how integrating the 4PADAFE instructional design methodology with generative artificial intelligence (GAI) tools helps develop innovative, pedagogically sound digital learning environments in higher education. To meet the demand for scalable and flexible instructional models, 4PADAFE offers a seven-phase, iterative framework [...] Read more.
This study investigates how integrating the 4PADAFE instructional design methodology with generative artificial intelligence (GAI) tools helps develop innovative, pedagogically sound digital learning environments in higher education. To meet the demand for scalable and flexible instructional models, 4PADAFE offers a seven-phase, iterative framework that connects pedagogical goals with the creative use of AI-powered tools. Using a qualitative exploratory approach, 20 Systems Engineering students applied the methodology to collaboratively create a four-week Massive Open Online Course (MOOC) titled “Generative Artificial Intelligence Tools for University Teaching.” They utilized ChatGPT, DALL·E, and Gamma to produce educational materials without direct input from subject-matter experts. Data collection included semi-structured interviews, non-participant observation, and analysis of student-created artifacts. The findings revealed increased learner autonomy, creativity, and digital skills, along with more efficient instructional design processes supported by prompt engineering and real-time feedback. The structured 4PADAFE framework helped participants align AI-generated content with specific learning outcomes while maintaining ethical safeguards. This study concludes that, with proper guidance and a systematic framework, students with technical backgrounds can serve as effective instructional designers, demonstrating the potential of combining structured methodologies and GAI to democratize high-quality course development in digital higher education. Full article
Show Figures

Figure 1

9 pages, 733 KB  
Brief Report
Narrative Medicine and AI in Anesthesiology Training: Teaching Empathy in End-of-Life Care
by Anna La Palma, Giuliana Scarpati, Giulia Savarese and Ornella Piazza
Int. Med. Educ. 2026, 5(1), 33; https://doi.org/10.3390/ime5010033 - 10 Mar 2026
Abstract
Teaching empathy remains a challenge in medical education, particularly in anesthesiology, where physicians frequently care for patients at the end of life. Narrative Medicine, centered on communicative competence and patients’ lived experience, offers a framework for cultivating reflective and relational skills. Meanwhile, artificial [...] Read more.
Teaching empathy remains a challenge in medical education, particularly in anesthesiology, where physicians frequently care for patients at the end of life. Narrative Medicine, centered on communicative competence and patients’ lived experience, offers a framework for cultivating reflective and relational skills. Meanwhile, artificial intelligence (AI) systems can generate expressions of empathy, raising questions about authentic moral engagement. To explore how narrative-based education, combined with AI-generated texts, may stimulate reflection, we implemented an exploratory narrative-based intervention involving 25 anesthesiology residents, supported by three tutors, integrating literature, film, and AI-generated narratives. After an introduction session, participants engaged with excerpts from the book What Are You Going Through and the film The Room Next Door, followed by reflective writing based on five prompts. The same prompts were submitted to ChatGPT (OpenAI, GPT-4o) for comparative analysis, discussed during a debriefing session. Reflective writings were assessed using an adapted REFLECT rubric, alongside qualitative lexical and semantic analyses. Most participants did not reach the highest levels of reflective capacity, while ChatGPT texts achieved higher REFLECT scores, primarily due to linguistic coherence. These findings suggest that empathic competence is neither automatically acquired through medical training nor reducible to verbal fluency. Rather, it requires structured training grounded in meaningful engagement with patients. Full article
Show Figures

Figure 1

38 pages, 2640 KB  
Article
Helpful or Harmful? Re-Evaluating Frugality in Retrieval-Augmented Generation for Medical Question Answering
by Richard Coric, Ebenezer F. Oloyede and Heriberto Cuayáhuitl
Mach. Learn. Knowl. Extr. 2026, 8(3), 64; https://doi.org/10.3390/make8030064 - 6 Mar 2026
Viewed by 141
Abstract
Medical question answering (QA) systems and conversational agents have attracted growing interest as tools that can assist clinicians, support medical students, and help patients navigate complex information sources. However, existing evaluations of retrieval strategies largely overlook the cost–benefit balance—here referred to as frugality, [...] Read more.
Medical question answering (QA) systems and conversational agents have attracted growing interest as tools that can assist clinicians, support medical students, and help patients navigate complex information sources. However, existing evaluations of retrieval strategies largely overlook the cost–benefit balance—here referred to as frugality, under realistic computational constraints. This work introduces a frugality-based evaluation framework that jointly assesses accuracy improvements and computational cost to determine when retrieval-augmented generation is beneficial in medical question answering, rather than evaluating retrieval effectiveness through accuracy alone. This study addresses these gaps through a systematic comparative framework that evaluates retrieval relevance, computational efficiency, and knowledge base composition across multiple biomedical QA tasks. We employ open-source LLMs (LlaMA-3-8B-Instruct, Mistral-7B-Instruct-v0.3, and DeepSeek-7B-Chat) across three benchmark medical QA datasets, including MedMCQA, MedQA-USMLE, and PubMedQA. In addition to that, we evaluate a dataset with larger contexts to simulate model distraction across the CliniQG4QA dataset using additional models (Meditron-7B, Qwen2.5-7B-Medical, Medgemma-4B, Phi-3-mini-4k-Instruct, and GPT4o-Mini). We examine how retrieval design choices alter the accuracy–latency trade-off, examining how relevance, corpus design, and hardware constraints interact in medical retrieval-augmented generation (RAG) systems. Our comprehensive results demonstrate when retrieval is genuinely beneficial versus when it imposes unnecessary computational costs, highlighting interactions between relevance and corpus designs in QA. Full article
Show Figures

Figure 1

12 pages, 1005 KB  
Article
Can Large Language Models Generate High-Quality Short-Answer Assessments? A Comparative Study in Undergraduate Medical Education
by Leo Morjaria, Levi Burns, Bhavya Gandhi, Keyna Bracken, Muhammad S. Farooq, Anthony J. Levinson, Quang Ngo and Matthew Sibbald
Appl. Sci. 2026, 16(5), 2535; https://doi.org/10.3390/app16052535 - 6 Mar 2026
Viewed by 166
Abstract
Background: Generative artificial intelligence (AI) tools including ChatGPT have the potential to augment the process of designing examinations and assessments for medical learners, leading to time and resource savings, and the ability to produce large volumes of practice problems tailored to learner-specific strengths [...] Read more.
Background: Generative artificial intelligence (AI) tools including ChatGPT have the potential to augment the process of designing examinations and assessments for medical learners, leading to time and resource savings, and the ability to produce large volumes of practice problems tailored to learner-specific strengths and weaknesses. Methods: This study compares the quality of free-text assessment problems and answer keys generated by ChatGPT to those produced by faculty educators for a renal and hematology curriculum subunit. Five expert reviewers reviewed a collection of 21 free-text assessment problems, 9 from a collection of historical assessment problems used in an undergraduate medical program and 12 produced with ChatGPT. Reviewers assigned a score from 1 to 5, reflecting the overall quality. Results: The average quality of problems generated by ChatGPT was greater than that of human-generated problems (4.00 vs. 2.71, p < 0.001). Using ordinal mixed-effect modeling, human-generated problems had significantly lower odds of receiving higher ratings than ChatGPT-generated problems (β = −2.43, 95% confidence interval −3.34 to −1.51, p < 0.001). Conclusions: It is suggested that ChatGPT can assist expert faculty educators in producing assessment tools, with direct benefits to medical learners, although it cannot entirely replace this role in its current state. Full article
Show Figures

Figure 1

24 pages, 6373 KB  
Article
Augmented Reality-Based Training System Using Multimodal Language Model for Context-Aware Guidance and Activity Recognition in Complex Machine Operations
by Waseem Ahmed and Qingjin Peng
Designs 2026, 10(2), 30; https://doi.org/10.3390/designs10020030 - 5 Mar 2026
Viewed by 169
Abstract
Augmented Reality (AR) and Large Language Models (LLMs) have made significant advances across many fields, opening new possibilities, particularly in complex machine operations. In complex operations, non-expert users often struggle to perform high-precision tasks and require constant supervision to execute tasks correctly. This [...] Read more.
Augmented Reality (AR) and Large Language Models (LLMs) have made significant advances across many fields, opening new possibilities, particularly in complex machine operations. In complex operations, non-expert users often struggle to perform high-precision tasks and require constant supervision to execute tasks correctly. This paper proposes a novel AR-MLLM-based training system that integrates AR, multimodal large language models (MLLMs), and prompt engineering to interpret real-time machine feedback and user activity. It converts extensive technical text into structured, step-by-step commands. The system uses a prompt structure developed through an iterative design method and refined across multiple machine operation scenarios, enabling ChatGPT to generate task-specific contextual digital overlays directly on the physical machines. A case study with participants was conducted to assess the effectiveness and usability of the AR-MLLM system in Coordinate Measuring Machine (CMM) operation training. The experimental results demonstrate high accuracy in task recognition and feature measurement activity. The data further show reduced time and user workload during task execution with the proposed AR-MLLM system. The proposed system not only provides real-time guidance and enhances efficiency in CMM operation training but also demonstrates the potential of the AR-MLLM design framework for broader industrial applications. Full article
Show Figures

Figure 1

11 pages, 763 KB  
Article
Descriptive Heterogeneity of the Hornblower Sign Across Scientific Literature, Search Engines, and Large Language Models
by Peter Melcher, Ralf Henkelmann, Susanne Schleif and Salim Youssef
Information 2026, 17(3), 258; https://doi.org/10.3390/info17030258 - 5 Mar 2026
Viewed by 133
Abstract
Background/Objectives: Digitalization of medical knowledge has improved access to information but also increased the spread of imprecise content. Repeated exposure to incorrect descriptions may lead to their normalization over time. This is particularly evident for the Hornblower sign, which is frequently conflated [...] Read more.
Background/Objectives: Digitalization of medical knowledge has improved access to information but also increased the spread of imprecise content. Repeated exposure to incorrect descriptions may lead to their normalization over time. This is particularly evident for the Hornblower sign, which is frequently conflated with the Patte test in the literature, online sources, and large language model outputs. This study systematically evaluates these descriptions and quantifies related inaccuracies. Methods: A three-step approach was applied to answer the question “What is the Hornblower sign?”. First, primary publications referring to the original description by Arthuis were analyzed. Second, the first 35 Google search results were systematically reviewed. Third, responses from five widely used LLMs (ChatGPT 5.1, Grok 4.1, Gemini 3 Pro, Perplexity, and DeepSeek-R1) were evaluated. All descriptions were assessed using a standardized 4-point scoring system (0–3 points) capturing content accuracy and correct differentiation between the Hornblower sign and the Patte test. Results: Fourteen original publications were included, yielding a mean score of 2.07. Correct descriptions were found in 50%, while 43% described only the Patte test. Among 34 evaluable Google search results, the mean score was 1.17, with 77% scoring ≤ 1 point. The five LLMs achieved a mean score of 1.8, demonstrating substantial variability and overall incomplete conceptual accuracy. Conclusions: Descriptions of the Hornblower sign show substantial heterogeneity and frequent inaccuracies across the scientific literature, online sources, and LLM outputs. Conflation with the Patte test undermines diagnostic reliability and limits study comparability. Critical source appraisal and adherence to original test descriptions are essential to maintain clinical and scientific rigor. Full article
(This article belongs to the Section Biomedical Information and Health)
Show Figures

Figure 1

16 pages, 1344 KB  
Review
Dr. Google vs. Dr. ChatGPT in Online Health Self-Consultation: A Scoping Review of Accuracy, Bias, and Actionability (2023–2025)
by Magdalena Trillo-Domínguez, Juan Ignacio Martin-Neira and María Dolores Olvera-Lobo
Informatics 2026, 13(3), 41; https://doi.org/10.3390/informatics13030041 - 5 Mar 2026
Viewed by 227
Abstract
The rapid adoption of generative artificial intelligence (AI) systems has transformed health information seeking, raising questions about their role as intermediaries in non-professional health self-consultation. This study compares Google Search and ChatGPT as paradigmatic models of algorithmic mediation of health information, focusing on [...] Read more.
The rapid adoption of generative artificial intelligence (AI) systems has transformed health information seeking, raising questions about their role as intermediaries in non-professional health self-consultation. This study compares Google Search and ChatGPT as paradigmatic models of algorithmic mediation of health information, focusing on accuracy, biases, information quality and potential harms. A scoping review was conducted following the PRISMA-ScR framework. Empirical studies published between 2023 and 2025 were retrieved from PubMed/MEDLINE, Web of Science (WoS) and Scopus. After screening and eligibility assessment, 63 original empirical studies were included. The results indicate that ChatGPT consistently outperforms Google Search in terms of factual accuracy and information quality, achieving moderate to high DISCERN scores (4–5 out of 5) and showing moderate to strong correlations with expert clinical evaluations. Users also tend to value ChatGPT responses positively due to their clarity, coherence and perceived empathy. However, these advantages coexist with significant structural limitations. Hallucinations are reported in an estimated 31–45% of references, source provenance remains opaque, linguistic complexity is high, and actionability is limited, with only around 40% of responses providing clearly actionable guidance. In contrast, Google Search offers greater source traceability and verifiability, but at the cost of fragmented information and higher exposure to commercial content. The review identifies critical research gaps related to behavioural impacts, critical health literacy, equity of access, professional integration and vulnerable contexts. Overall, the findings highlight the need for hybrid human–AI models, professional mediation and critical AI literacy to ensure safe, equitable and trustworthy use of generative AI in public health communication. Full article
Show Figures

Figure 1

7 pages, 332 KB  
Brief Report
Large Language Models (LLM) for Emergency Department Triage Based on Vital Signs
by Thomas G. Lederer, William C. Herring, Lama A. Ammar, Benjamin S. Abella, Donald J. Apakama, Ethan E. Abbott and Aditya C. Shekhar
Emerg. Care Med. 2026, 3(1), 9; https://doi.org/10.3390/ecm3010009 - 5 Mar 2026
Viewed by 125
Abstract
Introduction: Large language models (LLMs) have proven effective in many different fields, including the allocation of scarce resources. Triage within emergency departments (ED) is a core process that ensures the sickest patients are seen in a timely manner. Relatively little research has examined [...] Read more.
Introduction: Large language models (LLMs) have proven effective in many different fields, including the allocation of scarce resources. Triage within emergency departments (ED) is a core process that ensures the sickest patients are seen in a timely manner. Relatively little research has examined the use of existing LLMs in the triage process. Methods: 12 widely available LLMs were provided with real-world patient triage vital sign data from an academic trauma center in a major metropolitan area. The LLMs were asked to assign a triage score to each patient based on this information alone. The deviation between each LLM triage score and the real-world triage score for each patient was calculated, and the absolute value of the deviation was calculated and then averaged across the entire dataset per LLM. The average absolute value of deviation (AAVD) could then be used to compare LLMs against each other. All LLMs were blinded to the real-world triage score and received no additional training or instruction. Results: The models with the highest concordance with real-world triage scores were Claude Sonnet 4.5 (AAVD: 0.37; 62.37% concordance), ChatGPT-5 Instant (AAVD: 0.39; 62.89% concordance), and Claude Opus 4.1 (AAVD: 0.40; 62.37% concordance). The least accurate models were Gemini 2.5 Flash (AAVD: 0.42; 43.81% concordance), ChatGPT-4o Mini (AAVD: 0.49; 45.36% concordance), and ChatGPT-o3 (AAVD: 0.48; 48.45% concordance). Conclusions: This study analyzes the ability of LLMs to triage emergency department patients based primarily on vital sign data. Certain LLMs demonstrated moderate concordance with real-world triage scores. LLMs may be able to synthesize objective vital sign data and provide a triage recommendation. Further study could involve clinical validation against patient outcomes. Full article
(This article belongs to the Special Issue Application of Artificial Intelligence in Emergency Care)
Show Figures

Figure 1

17 pages, 1223 KB  
Article
Factors Driving Study Efficiency Gains and Exam Readiness from ChatGPT Use Among STEM Students: A Machine Learning Analysis
by Vishnu Kumar
Knowledge 2026, 6(1), 7; https://doi.org/10.3390/knowledge6010007 - 4 Mar 2026
Viewed by 100
Abstract
This study examines the factors driving perceived Study Efficiency and Exam Readiness associated with ChatGPT use among STEM students in higher education. Although prior research on generative artificial intelligence (GenAI) has largely focused on adoption and attitudes using descriptive or linear statistical approaches, [...] Read more.
This study examines the factors driving perceived Study Efficiency and Exam Readiness associated with ChatGPT use among STEM students in higher education. Although prior research on generative artificial intelligence (GenAI) has largely focused on adoption and attitudes using descriptive or linear statistical approaches, limited empirical work has explored how students’ interactions with such tools relate to learning-related outcomes. To address this gap, this study applies an interpretable machine learning (ML) framework to identify key predictors of learning gains from ChatGPT use. Data were obtained from a large-scale global survey of STEM students (n = 10,525) across 109 countries and territories, capturing usage patterns, perceived capabilities, satisfaction, and academic outcomes. Two eXtreme Gradient Boosting (XGBoost)-based ML classification models were developed to predict Study Efficiency and Exam Readiness, and SHapley Additive exPlanations (SHAP) were used to interpret feature-level contributions. The models achieved strong predictive performance for the high-gain class, with an accuracy of 0.93 (F1 = 0.96) for Study Efficiency and 0.86 (F1 = 0.92) for Exam Readiness. Results indicate that motivation, personalized learning support, improved access to knowledge, facilitation of study activities, and exam-focused study assistance are key predictors of learning gains. These findings offer empirical and practical insights for educators and policymakers seeking to design effective and pedagogically sound AI-assisted learning environments in STEM education. Full article
(This article belongs to the Special Issue Knowledge Management in Learning and Education)
Show Figures

Figure 1

3 pages, 144 KB  
Editorial
Learning to Live with Gen-AI
by Antony Bryant
Informatics 2026, 13(3), 38; https://doi.org/10.3390/informatics13030038 - 4 Mar 2026
Viewed by 152
Abstract
In 2023, in the wake of the launch of ChatGPT, based on GPT-3, we invited contributions on the Topic AI chatbots: threat or opportunity [...] Full article
Back to TopTop