Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

Search Results (162)

Search Parameters:
Keywords = embodied agents

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
24 pages, 1517 KB  
Article
The “Invisible” Heritage of Women in NeSpoon’s Lace Murals: A Symbolic and Educational Three-Case Study
by Elżbieta Perzycka-Borowska, Lidia Marek, Kalina Kukielko and Anna Watola
Arts 2025, 14(6), 129; https://doi.org/10.3390/arts14060129 (registering DOI) - 27 Oct 2025
Abstract
Street art increasingly reshapes aesthetic hierarchies by introducing previously marginalised media into the public sphere. A compelling example is the artistic practice of the Polish artist NeSpoon (Elżbieta Dymna), whose work merges the visual language of traditional lace with the communicative strategies of [...] Read more.
Street art increasingly reshapes aesthetic hierarchies by introducing previously marginalised media into the public sphere. A compelling example is the artistic practice of the Polish artist NeSpoon (Elżbieta Dymna), whose work merges the visual language of traditional lace with the communicative strategies of contemporary urban art. Active since the late 2000s, NeSpoon combines stencils, ceramic lace imprints, and large-scale murals to translate the intimacy of handcraft into the visibility of public space. Her works function as both aesthetic interventions and acts of civic pedagogy. This study employs a qualitative visual research design combining multi-site digital inquiry, iconological and semiotic analysis, and mini focus group (N = 22). Three purposefully selected cases: Łódź, Belorado, and Fundão, were examined to capture the site-specific and cultural variability of lace murals across Europe. The analysis demonstrates that lace functions as an agent of cultural negotiation and a medium of heritage literacy, understood here as embodied and place-based learning. In Łódź, it monumentalises textile memory and women’s labour embedded in the city’s industrial palimpsest. In Belorado, micro-scale responsiveness operates, strengthening the local semiosphere. In Fundão, lace enters an intermedial dialogue with azulejos, negotiating the boundary between craft and art while expanding local visual grammars. The study introduces the conceptualisation of the monumentalisation of intimacy in public art and frames heritage literacy as an embodied, dialogic, and community-oriented educational practice. Its implications extend to feminist art history, place-based pedagogy, urban cultural policy, and the preventive conservation of murals. The research elucidates how domestic craft once confined to the private interior operates in public space as a medium of memory, care, and inclusive aesthetics. Full article
(This article belongs to the Section Visual Arts)
Show Figures

Figure 1

31 pages, 5190 KB  
Article
MDF-YOLO: A Hölder-Based Regularity-Guided Multi-Domain Fusion Detection Model for Indoor Objects
by Fengkai Luan, Jiaxing Yang and Hu Zhang
Fractal Fract. 2025, 9(10), 673; https://doi.org/10.3390/fractalfract9100673 - 18 Oct 2025
Viewed by 250
Abstract
With the rise of embodied agents and indoor service robots, object detection has become a critical component supporting semantic mapping, path planning, and human–robot interaction. However, indoor scenes often face challenges such as severe occlusion, large-scale variations, small and densely packed objects, and [...] Read more.
With the rise of embodied agents and indoor service robots, object detection has become a critical component supporting semantic mapping, path planning, and human–robot interaction. However, indoor scenes often face challenges such as severe occlusion, large-scale variations, small and densely packed objects, and complex textures, making existing methods struggle in terms of both robustness and accuracy. This paper proposes MDF-YOLO, a multi-domain fusion detection framework based on Hölder regularity guidance. In the backbone, neck, and feature recovery stages, the framework introduces the CrossGrid Memory Block, Hölder-Based Regularity Guidance–Hierarchical Context Aggregation module, and Frequency-Guided Residual Block, achieving complementary feature modeling across the state space, spatial domain, and frequency domain. In particular, the HG-HCA module uses the Hölder regularity map as a guiding signal to balance the dynamic equilibrium between the macro and micro paths, thus achieving adaptive coordination between global consistency and local discriminability. Experimental results show that MDF-YOLO significantly outperforms mainstream detectors in metrics such as mAP@0.5, mAP@0.75, and mAP@0.5:0.95, achieving values of 0.7158, 0.6117, and 0.5814, respectively, while maintaining near real-time inference efficiency in terms of FPS and latency. Ablation studies further validate the independent and synergistic contributions of CGMB, HG-HCA, and FGRB in improving small-object detection, occlusion handling, and cross-scale robustness. This study demonstrates the potential of Hölder regularity and multi-domain fusion modeling in object detection, offering new insights for efficient visual modeling in complex indoor environments. Full article
Show Figures

Figure 1

51 pages, 4751 KB  
Review
Large Language Models and 3D Vision for Intelligent Robotic Perception and Autonomy
by Vinit Mehta, Charu Sharma and Karthick Thiyagarajan
Sensors 2025, 25(20), 6394; https://doi.org/10.3390/s25206394 - 16 Oct 2025
Viewed by 421
Abstract
With the rapid advancement of artificial intelligence and robotics, the integration of Large Language Models (LLMs) with 3D vision is emerging as a transformative approach to enhancing robotic sensing technologies. This convergence enables machines to perceive, reason, and interact with complex environments through [...] Read more.
With the rapid advancement of artificial intelligence and robotics, the integration of Large Language Models (LLMs) with 3D vision is emerging as a transformative approach to enhancing robotic sensing technologies. This convergence enables machines to perceive, reason, and interact with complex environments through natural language and spatial understanding, bridging the gap between linguistic intelligence and spatial perception. This review provides a comprehensive analysis of state-of-the-art methodologies, applications, and challenges at the intersection of LLMs and 3D vision, with a focus on next-generation robotic sensing technologies. We first introduce the foundational principles of LLMs and 3D data representations, followed by an in-depth examination of 3D sensing technologies critical for robotics. The review then explores key advancements in scene understanding, text-to-3D generation, object grounding, and embodied agents, highlighting cutting-edge techniques such as zero-shot 3D segmentation, dynamic scene synthesis, and language-guided manipulation. Furthermore, we discuss multimodal LLMs that integrate 3D data with touch, auditory, and thermal inputs, enhancing environmental comprehension and robotic decision-making. To support future research, we catalog benchmark datasets and evaluation metrics tailored for 3D-language and vision tasks. Finally, we identify key challenges and future research directions, including adaptive model architectures, enhanced cross-modal alignment, and real-time processing capabilities, which pave the way for more intelligent, context-aware, and autonomous robotic sensing systems. Full article
(This article belongs to the Special Issue Advanced Sensors and AI Integration for Human–Robot Teaming)
Show Figures

Figure 1

21 pages, 5726 KB  
Article
Embodied and Shared Self-Regulation Through Computational Thinking Among Preschoolers
by X. Christine Wang, Grace Yaxin Xing and Virginia J. Flood
Educ. Sci. 2025, 15(10), 1346; https://doi.org/10.3390/educsci15101346 - 11 Oct 2025
Viewed by 463
Abstract
While existing research highlights a positive association between computational thinking (CT) and self-regulation (SR) skills, limited attention has been given to the embodied and social processes within CT activities that support young children’s executive functions (EFs)—key components of SR. This study investigates how [...] Read more.
While existing research highlights a positive association between computational thinking (CT) and self-regulation (SR) skills, limited attention has been given to the embodied and social processes within CT activities that support young children’s executive functions (EFs)—key components of SR. This study investigates how preschoolers develop basic and higher-order EFs, such as focused attention, inhibitory control, causal reasoning, and problem-solving, through their engagement with a tangible programming toy in teacher-guided small groups in a university-affiliated preschool. Informed by a we-syntonicity framework that integrates Papert’s concepts of body/ego syntonicity and Schutz’s “we-relationship”, we conducted a multimodal microanalysis of video-recorded group sessions. Our analysis focuses on two sessions, the “Obstacle Challenge” and “Conditionals”, featuring four excerpts. Findings reveal that children leverage bodily knowledge and empathy toward the toy—named Rapunzel—to sustain attention, manage impulses, reason about cause-effect, and collaborate on problem-solving. Three agents shape these processes: the toy, fostering collective engagement; the teacher, scaffolding learning and emotional regulation; and the children, coordinating actions and sharing affective responses. These findings challenge traditional views of SR as an individual cognitive activity, framing it instead as an embodied, social, and situated practice. This study underscores the importance of collaborative CT activities in fostering SR during early childhood. Full article
(This article belongs to the Special Issue Computational Thinking and Programming in Early Childhood Education)
Show Figures

Figure 1

12 pages, 212 KB  
Entry
Sensing, Feeling, and Origins of Cognition
by Gordana Dodig-Crnkovic
Encyclopedia 2025, 5(4), 160; https://doi.org/10.3390/encyclopedia5040160 - 8 Oct 2025
Viewed by 344
Definition
Cognition is often modeled in terms of abstract reasoning and neural computation, yet a growing body of theoretical and experimental work suggests that the roots of cognition lie in fundamental embodied regulatory processes. This article presents a theory of cognition grounded in sensing, [...] Read more.
Cognition is often modeled in terms of abstract reasoning and neural computation, yet a growing body of theoretical and experimental work suggests that the roots of cognition lie in fundamental embodied regulatory processes. This article presents a theory of cognition grounded in sensing, feeling, and affect—capacities that precede neural systems and are observable in even the simplest living organisms. Based on the info-computational framework, this entry outlines how cognition and proto-subjectivity co-emerge in biological systems. Embodied appraisal—the system’s ability to evaluate internal and external conditions in terms of valence (positive/negative; good/bad)—and the capacity to regulate accordingly are described as mutually constitutive processes observable at the cellular level. This concept reframes cognition not as abstract symbolic reasoning but as value-sensitive, embodied information dynamics resulting from self-regulating engagement with the environment that spans scales from unicellular organisms to complex animals. In this context, information is physically instantiated, and computation is the dynamic, self-modifying process by which organisms regulate and organize themselves. Cognition thus emerges from the dynamic coupling of sensing, internal evaluation, and adaptive morphological (material shape-based) activity. Grounded in findings from developmental biology, bioelectric signaling, morphological computation, and basal cognition, this account situates intelligence as an affect-driven regulatory capacity intrinsic to biological life. While focused on biological systems, this framework also offers conceptual insights for developing more adaptive and embodied forms of artificial intelligence. Future experiments with minimal living systems or synthetic agents may help operationalize and test the proposed mechanisms of proto-subjectivity and affect regulation. Full article
(This article belongs to the Section Biology & Life Sciences)
16 pages, 2069 KB  
Article
“Can I Use My Leg Too?” Dancing with Uncertainty: Exploring Probabilistic Thinking Through Embodied Learning in a Jerusalem Art High School Classroom
by Dafna Efron and Alik Palatnik
Educ. Sci. 2025, 15(9), 1248; https://doi.org/10.3390/educsci15091248 - 18 Sep 2025
Viewed by 356
Abstract
Despite increased interest in embodied learning, the role of sensorimotor activity in shaping students’ probabilistic reasoning remains underexplored. This design-based study examines how high school students develop key probabilistic concepts, including sample space, certainty, and event probability, through whole-body movement activities situated in [...] Read more.
Despite increased interest in embodied learning, the role of sensorimotor activity in shaping students’ probabilistic reasoning remains underexplored. This design-based study examines how high school students develop key probabilistic concepts, including sample space, certainty, and event probability, through whole-body movement activities situated in an authentic classroom setting. Grounded in embodied cognition theory, we introduce a two-axis interpretive framework. One axis spans sensorimotor exploration and formal reasoning, drawing from established continuums in the literature. The second axis, derived inductively from our analysis, contrasts engagement with distraction, foregrounding the affective and attentional dimensions of embodied participation. Students engaged in structured yet open-ended movement sequences that elicited intuitive insights. This approach, epitomized by one student’s spontaneous question, “Can I use my leg too?”, captures the agentive and improvisational character of the embodied learning environment. Through five analyzed classroom episodes, we trace how students shifted between bodily exploration and formalization, often through nonlinear trajectories shaped by play, uncertainty, and emotionally driven reflection. While moments of insight emerged organically, they were also fragile, as they were affected by ambiguity and the difficulty in translating physical actions into mathematical language. Our findings underscore the pedagogical potential of embodied design for probabilistic learning while also highlighting the need for responsive teaching that balances structure with improvisation and supports affective integration throughout the learning process. Full article
Show Figures

Figure 1

21 pages, 1662 KB  
Article
Controllable Speech-Driven Gesture Generation with Selective Activation of Weakly Supervised Controls
by Karlo Crnek and Matej Rojc
Appl. Sci. 2025, 15(17), 9467; https://doi.org/10.3390/app15179467 - 28 Aug 2025
Viewed by 609
Abstract
Generating realistic and contextually appropriate gestures is crucial for creating engaging embodied conversational agents. Although speech is the primary input for gesture generation, adding controls like gesture velocity, hand height, and emotion is essential for generating more natural, human-like gestures. However, current approaches [...] Read more.
Generating realistic and contextually appropriate gestures is crucial for creating engaging embodied conversational agents. Although speech is the primary input for gesture generation, adding controls like gesture velocity, hand height, and emotion is essential for generating more natural, human-like gestures. However, current approaches to controllable gesture generation often utilize a limited number of control parameters and lack the ability to activate/deactivate them selectively. Therefore, in this work, we propose the Cont-Gest model, a Transformer-based gesture generation model that enables selective control activation through masked training and a control fusion strategy. Furthermore, to better support the development of such models, we propose a novel evaluation-driven development (EDD) workflow, which combines several iterative tasks: automatic control signal extraction, control specification, visual (subjective) feedback, and objective evaluation. This workflow enables continuous monitoring of model performance and facilitates iterative refinement through feedback-driven development cycles. For objective evaluation, we are using the validated Kinetic–Hellinger distance, an objective metric that correlates strongly with the human perception of gesture quality. We evaluated multiple model configurations and control dynamics strategies within the proposed workflow. Experimental results show that Feature-wise Linear Modulation (FiLM) conditioning, combined with single-mask training and voice activity scaling, achieves the best balance between gesture quality and adherence to control inputs. Full article
(This article belongs to the Section Computing and Artificial Intelligence)
Show Figures

Figure 1

7 pages, 171 KB  
Proceeding Paper
The Evolution of Intelligence from Active Matter to Complex Intelligent Systems via Agent-Based Autopoiesis
by Gordana Dodig-Crnkovic
Proceedings 2025, 126(1), 2; https://doi.org/10.3390/proceedings2025126002 - 18 Aug 2025
Viewed by 754
Abstract
Intelligence is a central topic in computing and philosophy, yet its origins and biological roots remain poorly understood. The framework proposed in this paper approaches intelligence as the complexification of agency across multiple levels of organization—from active matter to symbolic and social systems. [...] Read more.
Intelligence is a central topic in computing and philosophy, yet its origins and biological roots remain poorly understood. The framework proposed in this paper approaches intelligence as the complexification of agency across multiple levels of organization—from active matter to symbolic and social systems. Agents gradually acquire the capacity to detect differences, regulate themselves, and sustain identity within dynamic environments. Grounded in autopoiesis, cognition is reframed as a recursive, embodied process sustaining life through self-construction. Intelligence evolves as a problem-solving capacity of increasing organizational complexity: from physical self-organization to collective and reflexive capabilities. The model integrates systems theory, cybernetics, enactivism, and computational approaches into a unified info-computational perspective. Full article
29 pages, 1386 KB  
Article
A Hybrid Zero Trust Deployment Model for Securing O-RAN Architecture in 6G Networks
by Max Hashem Eiza, Brian Akwirry, Alessandro Raschella, Michael Mackay and Mukesh Kumar Maheshwari
Future Internet 2025, 17(8), 372; https://doi.org/10.3390/fi17080372 - 18 Aug 2025
Viewed by 737
Abstract
The evolution toward sixth generation (6G) wireless networks promises higher performance, greater flexibility, and enhanced intelligence. However, it also introduces a substantially enlarged attack surface driven by open, disaggregated, and multi-vendor Open RAN (O-RAN) architectures that will be utilised in 6G networks. This [...] Read more.
The evolution toward sixth generation (6G) wireless networks promises higher performance, greater flexibility, and enhanced intelligence. However, it also introduces a substantially enlarged attack surface driven by open, disaggregated, and multi-vendor Open RAN (O-RAN) architectures that will be utilised in 6G networks. This paper addresses the urgent need for a practical Zero Trust (ZT) deployment model tailored to O-RAN specification. To do so, we introduce a novel hybrid ZT deployment model that establishes the trusted foundation for AI/ML-driven security in O-RAN, integrating macro-level enclave segmentation with micro-level application sandboxing for xApps/rApps. In our model, the Policy Decision Point (PDP) centrally manages dynamic policies, while distributed Policy Enforcement Points (PEPs) reside in logical enclaves, agents, and gateways to enable per-session, least-privilege access control across all O-RAN interfaces. We demonstrate feasibility via a Proof of Concept (PoC) implemented with Kubernetes and Istio and based on the NIST Policy Machine (PM). The PoC illustrates how pods can represent enclaves and sidecar proxies can embody combined agent/gateway functions. Performance discussion indicates that enclave-based deployment adds 1–10 ms of additional per-connection latency while CPU/memory overhead from running a sidecar proxy per enclave is approximately 5–10% extra utilisation, with each proxy consuming roughly 100–200 MB of RAM. Full article
(This article belongs to the Special Issue Secure and Trustworthy Next Generation O-RAN Optimisation)
Show Figures

Figure 1

19 pages, 1635 KB  
Article
Integrating AI-Driven Wearable Metaverse Technologies into Ubiquitous Blended Learning: A Framework Based on Embodied Interaction and Multi-Agent Collaboration
by Jiaqi Xu, Xuesong Zhai, Nian-Shing Chen, Usman Ghani, Andreja Istenic and Junyi Xin
Educ. Sci. 2025, 15(7), 900; https://doi.org/10.3390/educsci15070900 - 15 Jul 2025
Viewed by 1362
Abstract
Ubiquitous blended learning, leveraging mobile devices, has democratized education by enabling autonomous and readily accessible knowledge acquisition. However, its reliance on traditional interfaces often limits learner immersion and meaningful interaction. The emergence of the wearable metaverse offers a compelling solution, promising enhanced multisensory [...] Read more.
Ubiquitous blended learning, leveraging mobile devices, has democratized education by enabling autonomous and readily accessible knowledge acquisition. However, its reliance on traditional interfaces often limits learner immersion and meaningful interaction. The emergence of the wearable metaverse offers a compelling solution, promising enhanced multisensory experiences and adaptable learning environments that transcend the constraints of conventional ubiquitous learning. This research proposes a novel framework for ubiquitous blended learning in the wearable metaverse, aiming to address critical challenges, such as multi-source data fusion, effective human–computer collaboration, and efficient rendering on resource-constrained wearable devices, through the integration of embodied interaction and multi-agent collaboration. This framework leverages a real-time multi-modal data analysis architecture, powered by the MobileNetV4 and xLSTM neural networks, to facilitate the dynamic understanding of the learner’s context and environment. Furthermore, we introduced a multi-agent interaction model, utilizing CrewAI and spatio-temporal graph neural networks, to orchestrate collaborative learning experiences and provide personalized guidance. Finally, we incorporated lightweight SLAM algorithms, augmented using visual perception techniques, to enable accurate spatial awareness and seamless navigation within the metaverse environment. This innovative framework aims to create immersive, scalable, and cost-effective learning spaces within the wearable metaverse. Full article
Show Figures

Figure 1

21 pages, 3136 KB  
Article
Negative Expressions by Social Robots and Their Effects on Persuasive Behaviors
by Chinenye Augustine Ajibo, Carlos Toshinori Ishi and Hiroshi Ishiguro
Electronics 2025, 14(13), 2667; https://doi.org/10.3390/electronics14132667 - 1 Jul 2025
Viewed by 1397
Abstract
The ability to effectively engineer robots with appropriate social behaviors that conform to acceptable social norms and with the potential to influence human behavior remains a challenging area in robotics. Given this, we sought to provide insights into “what can be considered a [...] Read more.
The ability to effectively engineer robots with appropriate social behaviors that conform to acceptable social norms and with the potential to influence human behavior remains a challenging area in robotics. Given this, we sought to provide insights into “what can be considered a socially appropriate and effective behavior for robots charged with enforcing social compliance of various magnitudes”. To this end, we investigate how social robots can be equipped with context-inspired persuasive behaviors for human–robot interaction. For this, we conducted three separate studies. In the first, we explored how the android robot “ERICA” can be furnished with negative persuasive behaviors using a video-based within-subjects design with N = 50 participants. Through a video-based experiment employing a mixed-subjects design with N = 98 participants, we investigated how the context of norm violation and individual user traits affected perceptions of the robot’s persuasive behaviors in the second study. Lastly, we investigated the effect of the robot’s appearance on the perception of its persuasive behaviors, considering two humanoids (ERICA and CommU) through a within-subjects design with N = 100 participants. Findings from these studies generally revealed that the robot could be equipped with appropriate and effective context-sensitive persuasive behaviors for human–robot interaction. Specifically, the more assertive behaviors (displeasure and anger) of the agent were found to be effective (p < 0.01) as a response to a situation of repeated violation after an initial positive persuasion. Additionally, the appropriateness of these behaviors was found to be influenced by the severity of the violation. Specifically, negative behaviors were preferred for persuasion in situations where the violation affects other people (p < 0.01), as in the COVID-19 adherence and smoking prohibition scenarios. Our results also revealed that the preference for the negative behaviors of the robots varied with users’ traits, specifically compliance awareness (CA), agreeableness (AG), and the robot’s embodiment. The current findings provide insights into how social agents can be equipped with appropriate and effective context-aware persuasive behaviors. It also suggests the relevance of a cognitive-based approach in designing social agents, particularly those deployed in sensitive social contexts. Full article
(This article belongs to the Special Issue Advancements in Robotics: Perception, Manipulation, and Interaction)
Show Figures

Graphical abstract

24 pages, 1658 KB  
Article
Modeling with Embodiment for Inquiry-Based Science Education
by Jordi Solbes, Rafael Palomar, M. Francisca Petit and Paula Tuzón
Educ. Sci. 2025, 15(7), 796; https://doi.org/10.3390/educsci15070796 - 20 Jun 2025
Viewed by 731
Abstract
Modeling is a fundamental scientific procedure for understanding nature, and it is also one of the basic strategies in inquiry-based science education. Among the various tools available for modeling, this article focuses on investigating a particular framework that uses embodiment to understand both [...] Read more.
Modeling is a fundamental scientific procedure for understanding nature, and it is also one of the basic strategies in inquiry-based science education. Among the various tools available for modeling, this article focuses on investigating a particular framework that uses embodiment to understand both macroscopic and microscopic phenomena. Within this approach, students actively engage as agents in the model and together build the final representation. For that, we present a specific methodology (the IBME approach) for inquiry-based modeling with embodiment. We specify the steps of the modeling approach, which were subsequently tested through instructional sequences based on this method with second-year students obtaining a degree in Primary Education at a public university. We analyzed the instructional sequences both quantitatively and descriptively. The quantitative analysis compares the results of an experimental group (n= 86) with a control group (n = 68) that does not work with inquiry-based modeling. Both groups address the same concepts, and at the end, they complete a questionnaire. The descriptive analysis discusses the details of the modeling process and the discussions that take place throughout the teaching sequences; on the other hand, it also summarizes the progress in the modeling process based on three qualitative parameters. The results obtained after implementing these sequences show significant differences compared to the control group. The descriptive analysis illustrates how students are able to reach the final model by inquiry, that is, through the discussion fostered by the modeling process itself, involving models of different levels of complexity. Full article
(This article belongs to the Special Issue Inquiry-Based Science Teaching and Learning)
Show Figures

Figure 1

34 pages, 1952 KB  
Article
Using Large Language Models to Embed Relational Cues in the Dialogue of Collaborating Digital Twins
by Sana Salman and Deborah Richards
Systems 2025, 13(5), 353; https://doi.org/10.3390/systems13050353 - 6 May 2025
Viewed by 1290
Abstract
Embodied Conversational Agents (ECAs) serve as digital twins (DTs), visually and behaviorally mirroring human counterparts in various roles, including healthcare coaching. While existing research primarily focuses on single-coach ECAs, our work explores the benefits of multi-coach virtual health sessions, where users engage with [...] Read more.
Embodied Conversational Agents (ECAs) serve as digital twins (DTs), visually and behaviorally mirroring human counterparts in various roles, including healthcare coaching. While existing research primarily focuses on single-coach ECAs, our work explores the benefits of multi-coach virtual health sessions, where users engage with specialized diet, physical, and cognitive coaches simultaneously. ECAs require verbal relational cues—such as empowerment, affirmation, and empathy—to foster user engagement and adherence. Our study integrates Generative AI to automate the embedding of these cues into coaching dialogues, ensuring the advice remains unchanged while enhancing delivery. We employ ChatGPT to generate empathetic and collaborative dialogues, comparing their effectiveness against manually crafted alternatives. Using three participant cohorts, we analyze user perception of the helpfulness of AI-generated versus human-generated relational cues. Additionally, we investigate whether AI-generated dialogues preserve the original advice’s semantics and whether human or automated validation better evaluates their lexical meaning. Our findings contribute to the automation of digital health coaching. Comparing ChatGPT- and human-generated dialogues for helpfulness, users rated human dialogues as more helpful, particularly for working alliance and affirmation cues, whereas AI-generated dialogues were equally effective for empowerment. By refining relational cues in AI-generated dialogues, this research paves the way for automated virtual health coaching solutions. Full article
Show Figures

Figure 1

28 pages, 5922 KB  
Article
Thoughtseeds: A Hierarchical and Agentic Framework for Investigating Thought Dynamics in Meditative States
by Prakash Chandra Kavi, Gorka Zamora-López, Daniel Ari Friedman and Gustavo Patow
Entropy 2025, 27(5), 459; https://doi.org/10.3390/e27050459 - 24 Apr 2025
Viewed by 1614
Abstract
The Thoughtseeds Framework introduces a novel computational approach to modeling thought dynamics in meditative states, conceptualizing thoughtseeds as dynamic attentional agents that integrate information. This hierarchical model, structured as nested Markov blankets, comprises three interconnected levels: (i) knowledge domains as information repositories, (ii) [...] Read more.
The Thoughtseeds Framework introduces a novel computational approach to modeling thought dynamics in meditative states, conceptualizing thoughtseeds as dynamic attentional agents that integrate information. This hierarchical model, structured as nested Markov blankets, comprises three interconnected levels: (i) knowledge domains as information repositories, (ii) the Thoughtseed Network where thoughtseeds compete, and (iii) meta-cognition regulating awareness. It simulates focused-attention Vipassana meditation via rule-based training informed by empirical neuroscience research on attentional stability and neural dynamics. Four states—breath_control, mind_wandering, meta_awareness, and redirect_breath—emerge organically from thoughtseed interactions, demonstrating self-organizing dynamics. Results indicate that experts sustain control dominance to reinforce focused attention, while novices exhibit frequent, prolonged mind_wandering episodes, reflecting beginner instability. Integrating Global Workspace Theory and the Intrinsic Ignition Framework, the model elucidates how thoughtseeds shape a unitary meditative experience through meta-awareness, balancing epistemic and pragmatic affordances via active inference. Synthesizing computational modeling with phenomenological insights, it provides an embodied perspective on cognitive state emergence and transitions, offering testable predictions about meditation skill development. The framework yields insights into attention regulation, meta-cognitive awareness, and meditation state emergence, establishing a versatile foundation for future research into diverse meditation practices (e.g., Open Monitoring, Non-Dual Awareness), cognitive development across the lifespan, and clinical applications in mindfulness-based interventions for attention disorders, advancing our understanding of the nature of mind and thought. Full article
(This article belongs to the Special Issue Integrated Information Theory and Consciousness II)
Show Figures

Figure 1

20 pages, 4055 KB  
Article
An Efficient Gaze Control System for Kiosk-Based Embodied Conversational Agents in Multi-Party Conversations
by Sunghun Jung, Junyeong Kum and Myungho Lee
Electronics 2025, 14(8), 1592; https://doi.org/10.3390/electronics14081592 - 15 Apr 2025
Viewed by 1077
Abstract
The adoption of kiosks in public spaces is steadily increasing, with a trend toward providing more natural user experiences through embodied conversational agents (ECAs). To achieve human-like interactions, ECAs should be able to appropriately gaze at the speaker. However, kiosks in public spaces [...] Read more.
The adoption of kiosks in public spaces is steadily increasing, with a trend toward providing more natural user experiences through embodied conversational agents (ECAs). To achieve human-like interactions, ECAs should be able to appropriately gaze at the speaker. However, kiosks in public spaces often face challenges, such as ambient noise and overlapping speech from multiple people, making it difficult to accurately identify the speaker and direct the ECA’s gaze accordingly. In this paper, we propose a lightweight gaze control system that is designed to operate effectively within the resource constraints of kiosks and the noisy conditions common in public spaces. We first developed a speaker detection model that identifies the active speaker in challenging noise conditions using only a single camera and microphone. The proposed model achieved a 91.6% mean Average Precision (mAP) in active speaker detection and a 0.6% improvement over the state-of-the-art lightweight model (Light ASD) (as evaluated on the noise-augmented AVA-Speaker Detection dataset), while maintaining real-time performance. Building on this, we developed a gaze control system for ECAs that detects the dominant speaker in a group and directs the ECA’s gaze toward them using an algorithm inspired by real human turn-taking behavior. To evaluate the system’s performance, we conducted a user study with 30 participants, comparing the system to a baseline condition (i.e., a fixed forward gaze) and a human-controlled gaze. The results showed statistically significant improvements in social/co-presence and gaze naturalness compared to the baseline, with no significant difference between the system and human-controlled gazes. This suggests that our system achieves a level of social presence and gaze naturalness comparable to a human-controlled gaze. The participants’ feedback, which indicated no clear distinction between human- and model-controlled conditions, further supports the effectiveness of our approach. Full article
(This article belongs to the Special Issue AI Synergy: Vision, Language, and Modality)
Show Figures

Figure 1

Back to TopTop