Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (342)

Search Parameters:
Keywords = large language model agent

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
39 pages, 5852 KB  
Article
SAPIENT: A Multi-Agent Framework for Corporate Reputation Intelligence Through Sentinel Monitoring and LLM-Based Synthetic Population Simulation
by Alper Ozpinar and Saha Baygul Ozpinar
Systems 2026, 14(4), 425; https://doi.org/10.3390/systems14040425 - 10 Apr 2026
Abstract
Corporate reputation teams rely on media monitoring and qualitative research, both limited in speed and coverage when digital narratives form rapidly. This paper proposes SAPIENT (Sentinel-Augmented Population Intelligence for Emerging Narrative Tracking), a multi-agent system that links a sentinel layer over public text [...] Read more.
Corporate reputation teams rely on media monitoring and qualitative research, both limited in speed and coverage when digital narratives form rapidly. This paper proposes SAPIENT (Sentinel-Augmented Population Intelligence for Emerging Narrative Tracking), a multi-agent system that links a sentinel layer over public text streams with a simulation layer that runs moderated, repeatable in silico focus-group sessions. The sentinel layer ingests social media, news, and forum text to produce a compact signal state (topics, sentiment, anomaly scores, risk labels), which conditions the simulation layer through an orchestrator. Persona agents and a moderator follow an Agentic Focus Group (AFG) protocol with repeated runs, variance reporting, and human review gates. We describe four sustainability communication scenarios: greenwashing backlash prediction, greenhushing risk assessment, campaign pre-testing, and crisis communication simulation. Nine experiments span 280 AFG runs across 20 conditions, three LLM backends (Claude Sonnet 4, GPT-4o, and Gemini 2.5 Flash), and a preregistered pilot human validation study with 54 participants. Signal conditioning improved simulation specificity (p=0.012). Cross-lingual sessions revealed a sentiment asymmetry between English and Turkish (p=0.001) with preserved persona rank ordering (r=0.81, p=0.015). Cross-model comparison showed consistent persona differentiation across all three backends (Pearson r>0.92, p<0.002 for all pairs). Sentiment was robust to prompt paraphrasing (p=0.061, n.s.), though credibility was sensitive to prompt wording (p<0.001). All significant results from Experiments 1–8 survived Benjamini–Hochberg correction. A preregistered pilot with 54 human participants on Prolific replicated the predicted credibility ranking across framing variants (p=0.004) but not the sentiment ranking, identifying a specific calibration target for future work. Full article
(This article belongs to the Section Artificial Intelligence and Digital Systems Engineering)
29 pages, 2439 KB  
Review
Agentic and LLM-Based Multimodal Anomaly Detection: Architectures, Challenges, and Prospects
by Mohammed Ayalew Belay, Amirshayan Haghipour, Adil Rasheed and Pierluigi Salvo Rossi
Sensors 2026, 26(8), 2330; https://doi.org/10.3390/s26082330 - 9 Apr 2026
Abstract
Anomaly detection is crucial in maintaining the safety, reliability, and optimal performance of complex systems across diverse domains, such as industrial manufacturing, cybersecurity, and autonomous systems. While conventional methods typically handle single data modalities, recently, there has been an increase in the application [...] Read more.
Anomaly detection is crucial in maintaining the safety, reliability, and optimal performance of complex systems across diverse domains, such as industrial manufacturing, cybersecurity, and autonomous systems. While conventional methods typically handle single data modalities, recently, there has been an increase in the application of multimodal detection in dynamic real-world environments. This paper presents a comprehensive review of recent research at the intersection of agentic artificial intelligence and large language-based multimodal anomaly detection. We systematically analyze and categorize existing studies based on the agent architecture, reasoning capabilities, tool integration, and modality scope. The main contribution of this work is a novel taxonomy that unifies agentic and multimodal anomaly detection methods, alongside benchmark datasets, evaluation methods, key challenges, and mitigation strategies. Furthermore, we identify major open issues, including data alignment, scalability, reliability, explainability, and evaluation standardization. Finally, we outline future research directions, with a particular emphasis on trustworthy autonomous agents, efficient multimodal fusion, human-in-the-loop systems, and real-world deployment in safety-critical applications. Full article
(This article belongs to the Special Issue Intelligent Sensors for Security and Attack Detection)
24 pages, 1584 KB  
Review
From Dialogue Systems to Autonomous Agents: A Modeling Framework for Ethical Generative AI in Healthcare
by James C. L. Chow and Kay Li
Information 2026, 17(4), 361; https://doi.org/10.3390/info17040361 - 9 Apr 2026
Abstract
The advancement of generative artificial intelligence (GAI) in healthcare is driving a transition from dialogue-based medical chatbots to workflow-embedded clinical AI agents. These agentic systems incorporate persistent state management, coordinated tool invocation, and bounded autonomy, enabling multi-step reasoning within institutional processes. As a [...] Read more.
The advancement of generative artificial intelligence (GAI) in healthcare is driving a transition from dialogue-based medical chatbots to workflow-embedded clinical AI agents. These agentic systems incorporate persistent state management, coordinated tool invocation, and bounded autonomy, enabling multi-step reasoning within institutional processes. As a result, traditional response-level evaluation frameworks are insufficient for understanding system behavior. This review provides a conceptual synthesis of the evolution from conversational systems to agentic architectures and proposes a system-level modeling framework for ethical clinical AI agents. We identify core architectural dimensions, including autonomy gradients, state persistence, tool orchestration, workflow coupling, and human–AI co-agency, and examine how these features reshape bias propagation pathways, error cascade dynamics, trust calibration, and accountability structures. Emphasizing that ethical risks emerge from longitudinal system interactions rather than isolated outputs, we argue for embedding fairness constraints, transparency mechanisms, and lifecycle governance directly within AI design. By outlining trajectory-level evaluation strategies, equity-aware development approaches, collaborative oversight models, and adaptive regulatory frameworks, this paper establishes a foundation for the responsible and trustworthy integration of agentic AI in healthcare. Full article
(This article belongs to the Special Issue Modeling in the Era of Generative AI)
Show Figures

Graphical abstract

20 pages, 1083 KB  
Article
FGeo-ISRL: A MCTS-Enhanced Deep Reinforcement Learning System for Plane Geometry Problem-Solving via Inverse Search
by Yang Li, Xiaokai Zhang, Cheng Qin, Zhengyu Hu and Tuo Leng
Symmetry 2026, 18(4), 628; https://doi.org/10.3390/sym18040628 - 9 Apr 2026
Abstract
Geometric problem-solving has always been a great challenge in the field of deductive reasoning and artificial intelligence. Symmetry is a defining characteristic of geometric shapes and properties. Consequently, the application of symmetry principles to geometric reasoning arises as a natural choice. To address [...] Read more.
Geometric problem-solving has always been a great challenge in the field of deductive reasoning and artificial intelligence. Symmetry is a defining characteristic of geometric shapes and properties. Consequently, the application of symmetry principles to geometric reasoning arises as a natural choice. To address the efficiency degradation and limited generalization, we propose FGeo-ISRL, a neural-symbolic inverse search framework whose core is the synergistic integration of a task-fine-tuned large language model and Monte Carlo Tree Search. Under the formal framework of FormalGeo, geometric theorems are iteratively applied starting from the given conditions and the target conclusion, in order to infer the necessary supporting premises. The large language model simultaneously serves as a policy network and a value network, guiding theorem application decisions and evaluating intermediate proof states, whereas the Monte Carlo Tree Search performs structured exploration over the state space, both training for policy refinement and inference for online search. The reinforcement learning agent is trained with a hybrid reward scheme, combining immediate feedback from the value difference and a sparse success reward. Experiments demonstrate the effectiveness and correctness of FGeo-ISRL. It not only achieves a Single-Step Theorem Accuracy of 90.2% and a Geometric Problem-Solving Accuracy of 83.14%, but also ensures that every step of the proof process remains readable, verifiable, and traceable. Full article
(This article belongs to the Section Computer)
Show Figures

Figure 1

37 pages, 10409 KB  
Article
A Scalable Framework for Street Interface Morphology Assessment via Automated Multimodal Large Language Model Agents
by Yuchen Wang, Yu Ye and Chao Weng
Land 2026, 15(4), 610; https://doi.org/10.3390/land15040610 - 8 Apr 2026
Abstract
Evaluating street interface morphology is essential for urban design, yet existing approaches often struggle to combine large-scale applicability with higher-level morphological interpretation. This study proposes a scalable framework for assessing street interface morphology using an automated multimodal large language model (MLLM) agent. Using [...] Read more.
Evaluating street interface morphology is essential for urban design, yet existing approaches often struggle to combine large-scale applicability with higher-level morphological interpretation. This study proposes a scalable framework for assessing street interface morphology using an automated multimodal large language model (MLLM) agent. Using street view imagery (SVI), the framework evaluates four core morphological dimensions—enclosure, continuity, transparency, and roughness–through two complementary analytical streams: objective geometric measurement and subjective morphological assessment. To support reliable evaluation, the framework incorporates a dual-benchmark strategy consisting of manually derived geometric measurements and expert-consensus ratings for calibration and validation. Applied in Shanghai, the framework demonstrated reliable performance across the evaluated dimensions. The optimized agent was further extended to continuous street-segment analysis, demonstrating its applicability to large-scale urban assessment. By integrating objective and subjective evaluation within a scalable and interpretable workflow, the proposed methodology provides a practical tool for street interface morphology analysis and urban design assessment. Full article
(This article belongs to the Section Land Planning and Landscape Architecture)
Show Figures

Figure 1

32 pages, 7135 KB  
Article
Evolutionary Multi-Objective Prompt Learning for Synthetic Text Data Generation with Black-Box Large Language Models
by Diego Pastrián, Nicolás Hidalgo, Víctor Reyes and Erika Rosas
Appl. Sci. 2026, 16(8), 3623; https://doi.org/10.3390/app16083623 - 8 Apr 2026
Abstract
High-quality training data are essential for the performance and generalization of artificial intelligence systems, particularly in dynamic environments such as adaptive stream processing for disaster response. However, constructing large and representative datasets remains costly and time-consuming, especially in domains where real data are [...] Read more.
High-quality training data are essential for the performance and generalization of artificial intelligence systems, particularly in dynamic environments such as adaptive stream processing for disaster response. However, constructing large and representative datasets remains costly and time-consuming, especially in domains where real data are scarce or difficult to obtain. Large Language Models (LLMs) provide powerful capabilities for synthetic text generation, yet the quality of generated data strongly depends on the design of input prompts. Prompt engineering is therefore critical, but it remains largely manual and difficult to scale, particularly in black-box settings where model internals are inaccessible. This work introduces EVOLMD-MO, a multi-objective evolutionary framework for automated prompt learning aimed at generating high-quality synthetic text datasets using black-box LLMs. The proposed approach formulates prompt optimization as a multi-objective search problem in which candidate prompts evolve through genetic operators guided by two complementary objectives: semantic fidelity to reference data and generative diversity of the produced samples. To support scalable optimization, the framework integrates a modular multi-agent architecture that decouples prompt evolution, LLM interaction, and evaluation mechanisms. The evolutionary process is implemented using the NSGA-II algorithm, enabling the discovery of diverse Pareto-optimal prompts that balance semantic preservation and diversity. Experimental evaluation using large-scale disaster-related social media data demonstrates that the proposed approach consistently improves prompt quality across generations while maintaining a stable trade-off between fidelity and diversity. Compared with a single-objective baseline, EVOLMD-MO explores a significantly broader semantic search space and produces more diverse yet semantically coherent synthetic datasets. These results indicate that multi-objective evolutionary prompt learning constitutes a promising strategy for black-box LLM-driven data generation, with potential applicability to adaptive data analytics and real-time decision-support systems in highly dynamic environments, pending broader validation across domains and models. Full article
(This article belongs to the Special Issue Resource Management for AI-Centric Computing Systems)
Show Figures

Figure 1

35 pages, 3162 KB  
Article
An LLM-Based Agentic Network Traffic Incident-Report Approach Towards Explainable-AI Network Defense
by Chia-Hong Chou, Arjun Sudheer and Younghee Park
J. Sens. Actuator Netw. 2026, 15(2), 32; https://doi.org/10.3390/jsan15020032 - 7 Apr 2026
Abstract
Traditional intrusion detection systems for IoT networks achieve high classification accuracy but lack interpretability and actionable incident-response capabilities, limiting their operational value in security-critical environments. This paper presents a graph-based multi-agent framework that integrates ensemble machine learning with Large Language Model (LLM)-powered incident [...] Read more.
Traditional intrusion detection systems for IoT networks achieve high classification accuracy but lack interpretability and actionable incident-response capabilities, limiting their operational value in security-critical environments. This paper presents a graph-based multi-agent framework that integrates ensemble machine learning with Large Language Model (LLM)-powered incident report generation via Retrieval-Augmented Generation (RAG). The system employs a three-phase architecture: (1) a lightweight Random Forest binary pre-detection, achieving 99.49% accuracy with a 6 MB model size for edge deployment; (2) ensemble classification combining Multi-Layer Perceptron, Random Forest, and XGBoost with soft voting and SHAP-based feature attribution for explainability; and (3) a ReAct-based summary agent that synthesizes classification results with external threat intelligence from Web search and scholarly databases to generate evidence-grounded incident reports. To address the challenge of evaluating non-deterministic LLM outputs, we introduce custom RAG evaluation metrics—faithfulness and groundedness implemented via the LLM-as-Judge framework. Experimental validation on the ACI IoT Network Dataset 2023 demonstrates ensemble accuracy exceeding 99.8% across 11 attack classes; perfect groundedness scores (1.0), indicating all generated claims derive from the retrieved context; and moderate faithfulness (0.64), reflecting appropriate analytical synthesis. The ensemble approach mitigates individual model weaknesses, improving the UDP Flood F1 score from 48% (MLP alone) to 95% through soft voting. This work bridges the gap between high-accuracy detection and trustworthy, actionable security analysis for automated incident-response systems. Full article
(This article belongs to the Special Issue Feature Papers in the Section of Network Security and Privacy)
Show Figures

Figure 1

32 pages, 1364 KB  
Article
XRL-LLM: Explainable Reinforcement Learning Framework for Voltage Control
by Shrenik Jadhav, Birva Sevak and Van-Hai Bui
Energies 2026, 19(7), 1789; https://doi.org/10.3390/en19071789 - 6 Apr 2026
Viewed by 232
Abstract
Reinforcement learning (RL) agents are increasingly deployed for voltage control in power distribution networks. However, their opaque decision-making creates a significant trust barrier, limiting their adoption in safety-sensitive operational settings. This paper presents XRL-LLM, a novel framework that generates natural language explanations for [...] Read more.
Reinforcement learning (RL) agents are increasingly deployed for voltage control in power distribution networks. However, their opaque decision-making creates a significant trust barrier, limiting their adoption in safety-sensitive operational settings. This paper presents XRL-LLM, a novel framework that generates natural language explanations for RL control decisions by combining game-theoretic feature attribution (KernelSHAP) with large language model (LLM) reasoning grounded in power systems domain knowledge. We deployed a Proximal Policy Optimization (PPO) agent on an IEEE 33-bus network to coordinate capacitor banks and on-load tap changers, successfully reducing voltage violations by 90.5% across diverse loading conditions. To make these decisions interpretable, KernelSHAP identifies the most influential state features. These features are then processed by a domain-context-engineered LLM prompt that explicitly encodes network topology, device specifications, and ANSI C84.1 voltage limits.Evaluated via G-Eval across 30 scenarios, XRL-LLM achieves an explanation quality score of 4.13/5. This represents a 33.7% improvement over template-based generation and a 67.9% improvement over raw SHAP outputs, delivering statistically significant gains in accuracy, actionability, and completeness (p<0.001, Cohen’s d values up to 4.07). Additionally, a physics-grounded counterfactual verification procedure, which perturbs the underlying power flow model, confirms a causal faithfulness of 0.81 under critical loading. Finally, five ablation studies yield three broader insights. First, structured domain context engineering produces synergistic quality gains that exceed any single knowledge component, demonstrating that prompt composition matters more than the choice of foundational model. Second, even an open source 8B-parameter model outperforms templates given the same prompt, confirming the framework’s backbone-agnostic value. Most importantly, counterfactual faithfulness increases alongside load severity, indicating that post hoc attributions are most reliable in the high-stakes regimes where trustworthy explanations matter most. Full article
Show Figures

Figure 1

37 pages, 1919 KB  
Article
LLMs for Integrated Business Intelligence: A Big Data-Driven Framework Integrating Marketing Optimization, Financial Performance, and Audit Quality
by Leonidas Theodorakopoulos, Aristeidis Karras, Alexandra Theodoropoulou and Christos Klavdianos
Big Data Cogn. Comput. 2026, 10(4), 110; https://doi.org/10.3390/bdcc10040110 - 5 Apr 2026
Viewed by 189
Abstract
Enterprise decision making in marketing, finance, and audit remains fragmented, leading to inefficient budget allocation and incomplete risk assessment. This study proposes an integrated, Big Data-driven decision-support framework that unifies Large Language Models (LLMs), attention-based marketing mix modeling, and multi-agent, game-theoretic optimization to [...] Read more.
Enterprise decision making in marketing, finance, and audit remains fragmented, leading to inefficient budget allocation and incomplete risk assessment. This study proposes an integrated, Big Data-driven decision-support framework that unifies Large Language Models (LLMs), attention-based marketing mix modeling, and multi-agent, game-theoretic optimization to coordinate cross-functional decisions. The architecture combines five modules: LLM-enhanced customer segmentation and customer lifetime value prediction, attention-weighted marketing mix modeling, multi-agent LLM systems for hierarchical budget optimization, attention-informed Markov multi-touch attribution, and LLM-augmented audit quality assessment. Empirical validation on a large-scale e-commerce dataset with 2.8 million customers and USD 156 million in marketing expenditure shows that marketing return on investment increases from 4.2 to 6.78 (61.4% relative improvement), financial forecasting error (MAPE) decreases from 12.8% to 4.7% (63.3% reduction), fraud detection accuracy improves by 29.8%, the Audit Quality Index reaches 0.951, and customer lifetime value prediction accuracy improves from 76.4% to 91.3%. By operationalizing the convergence of LLMs, attention mechanisms, and game-theoretic reasoning within a unified and empirically validated framework, the study delivers both theoretical advances and practically deployable tools for integrated business intelligence in digital economies. Full article
(This article belongs to the Section Large Language Models and Embodied Intelligence)
Show Figures

Figure 1

22 pages, 2152 KB  
Article
HCEA: A Multi-Agent Framework for Sustainable Human-Centered Entrepreneurship Based on a Large Language Model
by Yu Gao, Yanji Piao and Dongzhe Xuan
Sustainability 2026, 18(7), 3554; https://doi.org/10.3390/su18073554 - 4 Apr 2026
Viewed by 263
Abstract
Human-centered entrepreneurship considers employee well-being and uses the Sustainable Development Goals as its fundamental pillars. However, existing research predominantly focuses on institutional interventions and fails to provide integrated intelligent solutions for tackling human–machine collaboration issues in the context of digital transformation. Large language [...] Read more.
Human-centered entrepreneurship considers employee well-being and uses the Sustainable Development Goals as its fundamental pillars. However, existing research predominantly focuses on institutional interventions and fails to provide integrated intelligent solutions for tackling human–machine collaboration issues in the context of digital transformation. Large language models (LLMs) offer potential for affective computing and personalized support, but face critical gaps in ethical governance, privacy protection, and real-time risk intervention in sensitive entrepreneurial contexts. Our proposed Human-Centered Entrepreneurial Intelligent Agent (HCEA) framework achieves the unified optimization of task utility, empathetic expression, and ethical security by integrating a large language model core fine-tuned via a multi-objective hybrid loss function and a cluster of task-specialized intelligent agents. HCEA integrates retrieval-enhanced generation to ensure suggestion accuracy, a hierarchical data governance system for sensitivity-based privacy protection, and an independent risk detection module for real-time intervention and referral. We build the framework by constructing a hybrid entrepreneurial dataset, design the multi-agent architecture of decision support, emotion understanding and ethical risk tracking, and empirically evaluate both comparisons and ablation experiments. The results demonstrate that HCEA outperforms five baseline models across six key metrics, including entrepreneurship guidance relevance, emotion recognition, and high-risk recall. This study contributes to the intersection of digital transformation and sustainable entrepreneurship by providing a technically feasible, ethically grounded intelligent framework that empowers enterprises to reconcile efficiency with human-centric values, advancing SDG 8 (decent work and economic growth) and SDG 9 (industry, innovation, and infrastructure). Full article
Show Figures

Figure 1

12 pages, 284 KB  
Article
LLM-Based Control for Simulated Physical Reasoning: Modular Evaluation in the NeurIPS Embodied Agent Interface Challenge
by Hilmi Demirhan and Wlodek Zadrozny
AI 2026, 7(4), 131; https://doi.org/10.3390/ai7040131 - 3 Apr 2026
Viewed by 254
Abstract
Benchmark-driven evaluation helps distinguish between planning quality and interface reliability when large language models are utilized for embodied reasoning in simulation. Our submission to the Embodied Agent Interface Challenge (EAI) is evaluated across four stages of the pipeline. These being goal interpretation, subgoal [...] Read more.
Benchmark-driven evaluation helps distinguish between planning quality and interface reliability when large language models are utilized for embodied reasoning in simulation. Our submission to the Embodied Agent Interface Challenge (EAI) is evaluated across four stages of the pipeline. These being goal interpretation, subgoal decomposition, action sequencing, and transition modeling. The tasks run in the BEHAVIOR and VirtualHome simulators, which use constrained action vocabularies, fixed-object inventories and symbolic state representations within a standard evaluation protocol. Our system accesses the OpenAI API using GPT-4.1 for BEHAVIOR, GPT-4.1-mini for VirtualHome, and GPT-5-mini in later exploratory experiments across both environments. The schemas for each task determine how the outputs are structured, and outputs are regenerated when they do not follow the specification. On the final public leaderboard, our system ranked eighteenth overall with a score of 57.92, achieving 68.88 on BEHAVIOR and 46.96 on VirtualHome. In this paper, we describe our approach and discuss what these observations suggest about the strengths and limitations of current language models when used for embodied reasoning. Full article
(This article belongs to the Special Issue Integrating Large Language Models into Robotic Autonomy)
Show Figures

Figure 1

18 pages, 3157 KB  
Article
MINDS: A Modular Multi-Agent Decision-Support Framework for Dynamic Strategic Mine Planning
by Ricardo Nunes, Nathalie Risso and Moe Momayez
Mining 2026, 6(2), 26; https://doi.org/10.3390/mining6020026 - 2 Apr 2026
Viewed by 167
Abstract
Strategic Mine Planning (SMP) creates the long-term economic baseline for mining operations, yet economic variability necessitates Dynamic Mine Planning (DMP) to rapidly stress-test those financial assumptions. Currently, this capability is hindered by fragmented software ecosystems that require manual data handoffs, slowing iteration and [...] Read more.
Strategic Mine Planning (SMP) creates the long-term economic baseline for mining operations, yet economic variability necessitates Dynamic Mine Planning (DMP) to rapidly stress-test those financial assumptions. Currently, this capability is hindered by fragmented software ecosystems that require manual data handoffs, slowing iteration and breaking the audit trail between market data and valuation models. While Generative AI affords an opportunity to automate these workflows, its adoption in the mining industry is stalled by concerns over data quality and the risk of uncritical acceptance of automated outputs. Addressing these challenges, this paper describes the Mine Intelligence and Decision Support (MINDS) framework. We present MINDS as a modular reference architecture that uses Large Language Model (LLM) agents to orchestrate the economic evaluation process while maintaining strict engineering oversight. The system integrates a conversational interface with a multi-agent assessment layer that acts as an adversarial review, assessing price assumptions against market intelligence before generating economic valuation scenarios. A proof-of-concept using the Marvin copper benchmark evaluates the framework, demonstrating automated request-to-report orchestration, execution stability with an average debate latency of 10.69 s and a transparent decision audit trail. These findings show that MINDS can systematize economic scenario analysis without sacrificing the governance and verification required for definitive feasibility studies. Full article
(This article belongs to the Special Issue Mine Automation and New Technologies, 2nd Edition)
Show Figures

Graphical abstract

32 pages, 2911 KB  
Article
End-to-End Personalization via Unifying LLM Agents and Graph Attention Networks for Entertainment Recommendation
by Danial Ebrat, Sepideh Ahmadian and Luis Rueda
Information 2026, 17(4), 344; https://doi.org/10.3390/info17040344 - 2 Apr 2026
Viewed by 407
Abstract
Recommender systems are central to helping users navigate the rapidly expanding entertainment ecosystem, yet achieving strong personalization with limited feedback while maintaining interpretability remains difficult, particularly under cold-start conditions and heterogeneous item metadata. This work presents an end-to-end hybrid recommendation framework that unifies [...] Read more.
Recommender systems are central to helping users navigate the rapidly expanding entertainment ecosystem, yet achieving strong personalization with limited feedback while maintaining interpretability remains difficult, particularly under cold-start conditions and heterogeneous item metadata. This work presents an end-to-end hybrid recommendation framework that unifies a Large Language Model (LLM) with Graph Attention Network (GAT)-based collaborative filtering to improve both ranking accuracy and explanation quality across movies, books, and music. LLM-based agents first transform raw metadata such as titles, genres, descriptions, and auxiliary attributes into semantically grounded user and item profiles, which are embedded and used as initial node features in a user–item bipartite graph processed by a GAT-based recommender. Model optimization relies on a hybrid objective combining Bayesian Personalized Ranking, cosine-similarity regularization, and robust negative sampling to better align semantic and collaborative signals. Finally, in the post-processing stage, an LLM-based agent re-ranks the GAT outputs using a proposed Hybrid Confidence-Weighted Binary Search Tree, and another LLM-based agent that produces natural-language justifications tailored to each user. Experiments on diverse benchmark datasets and extensive ablations demonstrate that the proposed methodology increases precision, recall, NDCG, and MAP across various values of K. In addition, the post processing step is especially effective in cold-start scenarios, consistently strengthening recommendation metrics and enhancing transparency at smaller values of K. Overall, integrating LLM-enriched representations with attention-based graph modeling enables more accurate and explainable entertainment recommendations. Full article
Show Figures

Figure 1

31 pages, 2539 KB  
Article
Design and Evaluation of an AI-Based Conversational Agent for Travel Agencies: Enhancing Training, Assistance, and Operational Efficiency
by Pablo Vicente-Martínez, Emilio Soria-Olivas, Inés Esteve-Mompó, Manuel Sánchez-Montañés, María Ángeles García Escrivà and Edu William-Secin
AI 2026, 7(4), 123; https://doi.org/10.3390/ai7040123 - 1 Apr 2026
Viewed by 509
Abstract
The tourism industry faces increasing pressure for agile, personalized services, yet travel agencies struggle with fragmented knowledge scattered across isolated systems and legacy formats. While Large Language Models (LLMs) are widely applied in customer-facing roles, their potential to enhance internal operational efficiency remains [...] Read more.
The tourism industry faces increasing pressure for agile, personalized services, yet travel agencies struggle with fragmented knowledge scattered across isolated systems and legacy formats. While Large Language Models (LLMs) are widely applied in customer-facing roles, their potential to enhance internal operational efficiency remains largely underexplored. This study presents the design and evaluation of an intelligent assistant specifically for travel agency operations, built upon a Retrieval-Augmented Generation (RAG) architecture using Gemini 2.0 Flash. The system integrates heterogeneous data sources, including structured product catalogs and unstructured documentation processed via Optical Character Recognition (OCR), into a unified interface comprising work assistance, interactive training, and evaluation modules. Results demonstrate information retrieval times not greater than 45 s, ensuring its daily usability, while maintaining 95% accuracy. Furthermore, the system democratizes tacit senior expertise and accelerates new employee onboarding. This research validates RAG architectures as a powerful solution to knowledge fragmentation, shifting the strategic AI focus from customer automation to employee empowerment and operational optimization. Full article
Show Figures

Figure 1

22 pages, 3401 KB  
Article
TACOS: Task Agnostic Coordinator of a Multi-Drone System
by Alessandro Nazzari, Roberto Rubinacci and Marco Lovera
Drones 2026, 10(4), 251; https://doi.org/10.3390/drones10040251 - 31 Mar 2026
Viewed by 186
Abstract
When a single pilot is responsible for managing a multi-drone system, the task may demand varying levels of autonomy, from direct control of individual UAVs to group-level coordination to fully autonomous swarm behaviors for accomplishing high-level tasks. Enabling such interaction requires a framework [...] Read more.
When a single pilot is responsible for managing a multi-drone system, the task may demand varying levels of autonomy, from direct control of individual UAVs to group-level coordination to fully autonomous swarm behaviors for accomplishing high-level tasks. Enabling such interaction requires a framework that supports multiple modes of shared autonomy. As language models continue to improve in reasoning and planning, they provide a natural foundation for such systems. In this paper, we present TACOS (Task-Agnostic COordinator of a multi-drone System), a unified framework that enables high-level natural language control of multi-UAV systems through a Large Language Model (LLM). TACOS integrates three key capabilities into a single architecture: a one-to-many natural language interface for intuitive user interaction, an intelligent coordinator for translating user intent into structured task plans, and an autonomous agent that executes plans and interacts with the real world. TACOS allows an LLM to interact with a library of executable APIs, bridging semantic reasoning with real-time multi-robot coordination. We demonstrate the system on a real-world multi-drone system and conduct an ablation study to assess the contribution of each module. Full article
Show Figures

Figure 1

Back to TopTop