Next Article in Journal
Machine Learning Models for Predicting Mortality in Hemodialysis Patients: A Systematic Review
Previous Article in Journal
Refined Consolidation Settlement Calculation Based on the Oedometer Tests for Normally and Overconsolidated Clays
Previous Article in Special Issue
Beyond Binary Dialogues: Research and Development of a Linguistically Nuanced Conversation Design for Social Robots in Group–Robot Interactions
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Retrieving Memory Content from a Cognitive Architecture by Impressions from Language Models for Use in a Social Robot

by
Thomas Sievers
* and
Nele Russwinkel
Institute of Information Systems, University of Lübeck, Ratzeburger Allee 160, 23562 Lübeck, Germany
*
Author to whom correspondence should be addressed.
Appl. Sci. 2025, 15(10), 5778; https://doi.org/10.3390/app15105778
Submission received: 19 April 2025 / Revised: 17 May 2025 / Accepted: 20 May 2025 / Published: 21 May 2025
(This article belongs to the Special Issue Advances in Cognitive Robotics and Control)

Abstract

:
Large Language Models (LLMs) and Vision-Language Models (VLMs) have the potential to significantly advance the development and application of cognitive architectures for human–robot interaction (HRI) to enable social robots with enhanced cognitive capabilities. An essential cognitive ability of humans is the use of memory. We investigate a way to create a social robot with a human-like memory and recollection based on cognitive processes for a better comprehensible and situational behavior of the robot. Using a combined system consisting of an Adaptive Control of Thought-Rational (ACT-R) model and a humanoid social robot, we show how recollections from the declarative memory of the ACT-R model can be retrieved using data obtained by the robot via an LLM or VLM, processed according to the procedural memory of the cognitive model and returned to the robot as instructions for action. Real-world data captured by the robot can be stored as memory chunks in the cognitive model and recalled, for example by means of associations. This opens up possibilities for using human-like judgment and decision-making capabilities inherent in cognitive architectures with social robots and practically offers opportunities of augmenting the prompt for LLM-driven utterances with content from declarative memory, thus keeping them more contextually relevant. We illustrate the use of such an approach in HRI scenarios with the social robot Pepper.

1. Introduction

The human brain, complex as it is as a result of unplanned evolution, stores memories in a rather disorganized way compared to the precise addressing system of computers [1,2]. It is organized according to context, and topics are interwoven in complex associative networks and are retrieved by clues and associations, not by pure addressing in a memory as with a computer. Human memory is superior to that of computers in certain respects, for example, through automatic prioritization with regard to rewards and frequently used elements [3,4]. However, memories that are unused soon decay into oblivion. Furthermore, unlike the addressing system of a computer, which is the result of planned engineering following the von Neumann architecture, the human way of thinking is susceptible to the formation of false recollections through incorrect associations.
However, perhaps the Cartesian mind–body dualism, which ignores the importance of brain-inspired embodiment for the development of human-like Artificial Cognition (ACo), leads to a dead end [5]. Personal experiences, partly mediated by social interaction and accessible to recollection, are essential for an Artificial Intelligence (AI) agent to develop cognitive abilities and learn about its environment. Such an agent should be able to imagine actions using mental representations before executing them and justify them using the cognitive ability of prospection, including planning, predicting, imagining scenarios, and possible future events. A cognitive architecture seems to be crucial in equipping the agent with such capabilities. In this work, we ask the question whether a social robot can be endowed with human-like memory capabilities by using a cognitive architecture for information retrieval and processing and a Large Language Model (LLM) that is also capable of vision. A key criterion here is the ability to retrieve recollections from memory using associated expressions and to prepare the ground for further human-like cognitive abilities. We show a technical method for implementation.
LLMs like ChatGPT amaze the world with their seemingly human-like abilities in some respects. They provide advanced reasoning capabilities, blending intuitive and deliberate cognitive processes [6]. With their sophisticated probabilistic model of language use to generate a semantically consistent text, they require a form of effective memory that has surprising similarities with central features of human memory [7]. LLMs also help robotic systems improve their generalization capabilities in dynamic and complex real-world environments and can significantly increase their behavior planning and execution capabilities, enabling robots to engage with their environment in a human-like manner [8].
Vision-Language Models (VLMs) are multimodal AI systems created by combining an LLM with a vision encoder that gives the LLM the ability to “see”. They provide assistance with complex tasks such as creating captions and answering visual questions [9]. VLMs are capable of performing a variety of tasks after learning relationships between images and language from large data sets, such as answering questions about images, finding sentences that correspond well with images, and finding image regions that correspond to texts. These skills can be used in many ways for robotics, for example, for robot movements, state recognition, object recognition, affordance recognition, relation recognition, and anomaly detection [10,11].
Cognitive architectures, on the other hand, refer both to a theory about the structure of the human mind and to a computer-based implementation of such a theory. They are particularly suitable for general problem solving in areas that are not clearly defined and for human cognitive modeling [12]. Cognitive architectures attempt to describe and integrate the basic mechanisms of human cognition. In doing so, they rely on empirically supported assumptions from cognitive psychology. Their formalized models can be used to react flexibly to actions in a human-like manner and—when used in a robot—to develop a situational understanding regarding human behavior for adequate reactions. We apply the well-known and successfully implemented cognitive architecture Adaptive Control of Thought-Rational (ACT-R) for our test scenarios [13]. ACT-R is a hybrid cognitive architecture with a symbolic and sub-symbolic structure. It is based on psychological principles and offers a comprehensive model of cognitive processing. ACT-R models run in real time and are robust to errors or unexpected events.
Intuitive decision making as a subjective, particularly a human type of decision making, is based on implicit knowledge that is transmitted to the conscious mind at the time of the decision through affect or unconscious cognition. Computational models of intuitive decision making can be expressed as instance-based learning using the ACT-R cognitive architecture [14]. In Instance-Based Learning Theory (IBLT), past experiences (i.e., instances) are retrieved using cognitive mechanisms of a cognitive architecture. IBLT proposes learning mechanisms related to a decision-making process, such as instance-based knowledge and recognition-based retrieval. These learning mechanisms can be implemented in an ACT-R model [15,16]. A social robot would benefit from such abilities, for example, to solve problems or develop a kind of intuition for the situation. With regard to Human–Robot Interaction (HRI), a combination of robot sensor technology and data processing with such an architecture offers the possibility of dealing with information from the robot’s real world in cognitive models. A cognitive architecture may also be used to add a “human component” to robotic applications, as the procedural processes of a mental model behave differently—more human-like—than conventional algorithms [17].
In HRI applications that use an LLM for speech generation and/or a VLM for image content recognition, a cognitive model can use its memory to provide facts and the context of a particular scenario to the language model, bringing in personalized experiences by recalling unique memories collected by the robot. With the help of prompt augmentation, conclusions of a mental model can be taken into account in instruction generation via prompts for the LLM and thus reduce the weaknesses of the language model, such as ignorance of the current individual facts, and thus, hallucination when relevant facts are unclear. Furthermore, language models are good at fast automatic reasoning, but less capable of high-level cognition to enable complex mental operations and “slow thinking” following the dual process theory of human cognition [18]. A combination of the advantages of both approaches would result in a system that performs involuntary reactions as automatic tasks (e.g., formulating a sentence) and also incorporates and takes into account more complex human-like cognitive abilities such as memory. This would open up the possibility of significantly increasing the capabilities of a social robot in terms of incorporating and assessing human actions and intentions.
The cognitive architecture of ACT-R comprises a declarative and a procedural memory, whereby the declarative memory supports lexical knowledge by encoding, storing, and retrieving semantic knowledge, as in humans, while the procedural memory enables the learning of habits and skills [19,20,21,22]. Using the ACT-R chunk and memory system to store, retrieve, and process facts and impressions of a real-world scenario and utilizing knowledge from the procedural memory of the cognitive model allows an LLM to incorporate these facts into utterances, paving the way for a more reliable, evidence-based, and human-like application of LLMs.
For testing our approach, we applied OpenAI’s Generative Pretrained Transformer (GPT) language model in a setting with a social robot in dialog with a human, where the robot used the GPT model to generate speech, processed the content of the dialog, and recognized image content from its front camera [23]. However, the proposed methods can also be used beyond a HRI context. As a cognitive architecture, we used the standalone application of ACT-R 7, for which we created a cognitive model in LISP [24]. The connection between the cognitive model and our robot application, which also contained the Application Programming Interface (API) to the LLM, was realized via a TCP/IP connection of the dispatcher, which acts as a server for client applications to ACT-R [25].
To incorporate different sensory abilities from the robot, we used the visual capabilities of the GPT model in one test method and text generation in another. In principle, the procedure remained the same with regard to the use of the cognitive model and its memory function, except that the LLM was used once to describe image content and once to describe textual content. In the vision method, the language model was instructed to analyze the content of images from the front camera on the robot’s head and describe the main content of each image in three keywords or key phrases. In the other case, the utterances of a person in dialog with the robot were processed by an LLM so that the core content of the human question or problem was also expressed in three keywords. The number of keywords used for our experiments was arbitrarily set to three in order to ensure a certain variability for the association possibilities via the keywords, and at the same time, to keep the number of required productions of the cognitive model manageable. The principle outlined here is not dependent on the number of keywords.
These keywords or phrases were passed as chunks to an ACT-R cognitive model, where they were processed with productions from procedural memory to search the declarative memory for existing memory content indexed by chunks in the same or similar form. The memory content contained an additional chunk that represented the actual recollection and could, for example, represent a fact worth remembering in the form of a short sentence. If there was a positive correlation between keyword chunks from the LLM and memory chunks, this recollection was passed to the robot application and thus to the LLM for prompt augmentation, which generated a response based on this knowledge. In principle, it is possible to accumulate factual knowledge a priori by creating corresponding chunks in declarative memory. In the following, we show a technical implementation of such an architecture as a combination of LLM/VLM integration, a humanoid social robot, and cognitive architecture and give examples of applications.

2. Related Work

Robots are able to use Large Multimodal Models (LMMs) to comprehend and execute tasks based on natural language input and environmental cues [26]. The integration of foundation models such as LLMs and VLMs can effectively improve robot intelligence [8,27]. VLMs help, for example, equip robots with the ability for physically based task planning [28]. The use of GPT-4V for analyzing videos of humans performing tasks to obtain textual explanations of environmental and action details, combined with a GPT-4-based task planner that encodes these details into a symbolic task plan for a robot, was presented by Wake et al. [29].
Yoshida et al. investigated possibilities for the development of a “minimal self” with a sense of agency and ownership in a robot that was able to mimic human movements and emotions by using human knowledge from language models [30]. They did not implement a model of a cognitive architecture but used GPT-4’s motion generation and image recognition capabilities. VLMs as a basis for metacognitive thinking can enable robots to understand and improve their own processes, avoid hardware failures, and thus increase their resilience [31].
Given the recent successes of language models, there has been an increased interest in the interplay between LLMs and cognitive architectures. Ideas for a combination of both are emerging, or the creation of a cognitive architecture based on an LLM is being considered. Niu et al. provided an overview of the similarities, differences, and challenges between LLMs and cognitive science by analyzing methods for assessing the potential cognitive capabilities of LLMs, discussing biases and limitations, and an integration of LLM with cognitive architectures [32]. In the novel neuro-symbolic architecture presented by Wu et al., human-centered decision making was enabled through the integration of ACT-R with LLMs by using knowledge from the decision process of the cognitive model as neural representations in trainable layers of the LLM [33]. This improved the ability for grounded decision making. Gonzáles-Santamarta et al. provided an example of integrating an LLM with a cognitive architecture to enable planning and reasoning in autonomous robots [34]. They showed how to use the reasoning capabilities of LLMs in the MERLIN2 cognitive architecture integrated in ROS 2.
Berto et al. investigated how humans and cognitive robots with different value systems and motivations understand each other’s needs in free interactions [35]. For this purpose, a cognitive architecture was developed that combines perception with internal motivation and an intrinsic value system for determining actions. They defined an autonomous cognitive agent as an agent that is able to autonomously predict the outcome of its actions by drawing on past memories, current perceptions of the environment and other agents, and anticipation of the behavior of both the environment and its own actions.
The development of robots that can help people with their everyday tasks is inspired by many fields, with a particular focus on psychological studies such as Theory of Mind (ToM), a cognitive ability that humans develop in childhood to infer their own mental states and the mental states of other people such as beliefs, desires, knowledge, feelings, and intentions. In their review of ToM for autonomous systems such as robots, Hellou et al. explored the use of probabilistic models using Bayesian inference, mentioning the need to combine such models with data-driven models including Deep Learning (DL) and Reinforcement Learning (RL) due to limitations in Bayesian Network (BN) generalization [36]. However, they also mentioned a lack of data regarding DL for ToM scenarios and data limitations in RL models, the latter of which could be overcome by simulations. Finally, they emphasized the need for further studies on the integration of such cognitive capabilities with machines in real-world scenarios.
In order to relate to people’s mental states, robots must be able to not only recognize them but also remember them in the context of the interactions. The use of a cognitive architecture connected to a social robot to store and process memory chunks from a language game between the robot and a human was demonstrated by Sievers et al. [37]. An ACT-R model received words from the robot, searched for matching associations already present in memory, and provided feedback to an LLM on these associations when appropriate. This way of grounding abstract words and ideas should help constrain possible meanings.
Knowles et al. proposed a system architecture that combined LLMs and cognitive architectures with an analogy to “fast” and “slow” thinking in human cognition [18,38]. Leivada et al. explored whether the current generation of LLMs is able to develop grounded cognition that incorporates prior expectations and prior world experiences to perceive the big picture [39]. Insights from human cognition and psychology anchored in cognitive architectures could contribute to the development of systems that are more powerful, reliable, and human-like [40]. This dual-process architecture and the hybrid neuro-symbolic approach to overcoming the limitations of current LLMs is seen as particularly important.
He et al. investigated the Long-Term Memory (LTM) capabilities of AI systems such as LLM-based agents by creating a mapping system between human LTM and AI LTM mechanisms and proposing an extension of current cognitive architectures for the creation of next-generation AI systems with LTM [41]. The significance of LTM for the foundation of AI self-evolution—allowing LLMs to evolve during inference—based on limited data or interactions versus large-scale training on increasingly large datasets was emphasized by Jiang et al. [42].
A human-like memory architecture to improve the cognitive abilities of LLM-based dialog agents was proposed by Hou et al. [43]. The proposed architecture enabled agents to autonomously retrieve recollections required to generate responses, thereby addressing limitations of the temporal awareness of LLMs. Memories from the interaction history with the user were stored in a database containing the content and temporal context of each reminder. This allowed the agent to understand and utilize specific memory content relevant to the user in a temporal context.
In their Cognitive Architectures for Language Agents (CoALA), Sumers et al. proposed a language agent with modular memory components, a structured action space for interacting with internal memory and external environments, and a generalized decision-making process for selecting actions [44]. Their framework draws from the rich history of symbolic artificial intelligence and cognitive science and combines some longstanding findings with current research on LLMs for tasks that require basic knowledge or reasoning. Furthermore, Retrieval-Augmented Generation (RAG) can offer a different approach to challenges such as hallucinations, outdated knowledge, and untraceable reasoning processes in LLMs [45,46]. RAG does not retrieve knowledge from unknown scenarios, but provides an approach to combine external sources in LLMs with many optimizations to incorporate the appropriate knowledge. The use of a cognitive architecture memory capable of associations with recollections of similar situations experienced should be able to react more accurately to comparable new scenarios and control the LLM in a more targeted manner. However, testing this is beyond the scope of this paper.
The production of semantically consistent text requires a form of effective memory, especially in individual scenarios. An examination of the memory properties of LLMs revealed similarities with important features of human memory, but in the case of an LLM, these are learned from the statistics of the training text data and not from the architecture of the LLM [7]. ACT-R/E was developed for building embodied models of people with robots [47]. However, apart from that, the use of architectures that are modeled on human cognitive behavior, which like ACT-R are based on cognitive psychological findings and have been tried and tested in the psychological field for many years, is to our knowledge largely unexplored for use in social robots. For a more general, meaningful approach and a more human-like way of retrieving memories from a recollection—which, is also explainable in terms of the accessible memory contents and thus poses no explanatory problems—we propose using the declarative memory of a cognitive architecture in conjunction with the procedural abilities of a cognitive model, as described below.

3. Methods

Our system consists of the humanoid social robot Pepper with a software application that connects to OpenAI’s GPT-4o model via an API, and it communicates over a wireless TCP/IP network with the current version of the ACT-R 7 cognitive architecture software running on a remote computer. ACT-R offers the technical possibility of integrating a cognitive model with bidirectional communication into an application of choice. To establish a remote connection from the robot application to ACT-R, we used its remote interface—the dispatcher [25]. Figure 1 shows the setup of the bidirectional connection between the dispatcher, which acts as a kind of server, and the client application of the robot.

3.1. Humanoid Social Robot Pepper

The humanoid social robot Pepper, shown in Figure 2, was developed by Aldebaran and first released in 2015 [48]. The robot is 120 centimeters tall and optimized for human interaction. It is able to interact with people through conversation, gestures, and its touch screen, and it is therefore well suited for studies on human–robot interaction.
Pepper is equipped with four directional microphones in their head and speakers for voice output. The robot features an open and fully programmable platform so that developers can program their own applications to run on Pepper using software development kits (SDKs) for programming languages like C++, Python, Java or Kotlin. Our robot application was programmed in Kotlin (221-1.8.0-release-for-android-studio-AS5591.52) using Android Studio Electric Eel | 2022.1.1 Patch 2 for a Pepper robot with Naoqi version: 2.9.5 and the Pepper SDK 1.5.3. In a normal dialog mode with humans, the robot application forwarded utterances of the human dialog partner as input to the OpenAI API, which returned a dictionary with the status and response of the API. With each API call, the entire dialog was transferred to the GPT model. This allows the model to constantly ‘remember’ what was previously said and refer to it as the dialog progresses, enabling a comprehensive exchange between human and robot.

3.2. Cognitive Model

In ACT-R, declarative knowledge is represented in the form of chunks, i.e., representations of individual properties, each of which can be accessed via a labeled slot. The cognitive model programmed in LISP for our test scenarios should receive chunks with keywords from the robot application and checking whether chunks for these keywords are already stored in the declarative memory of the model in a suitable combination. We assumed the transfer of three keywords. The programmed productions of the procedural memory checked all combinations of the sequence of keywords for a match with memory content and generated a hit for two out of three or for all three matching keywords. In this case, the associated memory content was called up and the content of a special slot of the memory content chunk from this recollection was returned to the robot application.
Our LISP code defined a chunk type for memory content in declarative memory as follows: chunk-type keyword, asso-one, asso-two, asso-three, phrase.
This chunk type ‘keyword’ featured three labeled slots ‘asso-one’, ‘asso-two’, and ‘asso-three’ as well as a ‘phrase’ slot, which stored the memory content we wanted to retrieve. For example, this phrase could be a remembered fact used to constrain a system prompt for the LLM that the robot used to talk to humans. Each chunk stored in the declarative memory also has an arbitrary name. Figure 3 shows the process of searching for a matching memory chunk. The cognitive model searches the declarative memory for suitable keywords from the chunks stored there. First, it checks whether there are chunks in which all three keywords match; otherwise, it checks whether two keywords match. The order of the keywords as created by the LLM does not have to match the keyword order in the chunks in the memory. Having found a corresponding memory chunk, the cognitive model returned the sentence stored in the ‘phrase’ slot of this chunk as a fact phrase.

Workflow for Retrieving Memory Facts Based on Keywords from the LLM

  • The LLM creates chunks with three keywords related to the task at hand in response to what a person has said in the interaction.
  • These keyword chunks are transferred to the ACT-R model.
  • The cognitive model searches for associated chunks in the declarative memory.
  • Having found a corresponding memory chunk, the cognitive model returns the content of the ‘phrase’ slot of this chunk as a fact phrase.
  • This fact phrase is transferred back to the robot application and used for the prompt augmentation of the LLM.
  • The LLM provides an answer that includes the remembered fact.
The following is an example of the possible content contained in such a memory chunk for a train station scenario assumed in one of our tests as follows: asso-one platform-1, asso-two location, asso-three direction-assistance, and phrase Platform 1 is the neighboring platform right here.

3.3. Basic Settings of the LLM

We chose GPT-4o to create the conversational parts of the robot. The OpenAI API provides various hyperparameters that can be used to control the model behavior during an API call. The value for temperature was set to 0 in order to obtain consistent responses and exclude any randomness as far as possible. Positive values of the presence penalty would penalize new tokens, depending on whether they already appear in the text, and increase the probability that the model will talk about a new topic. We set this value to zero. Positive values of the frequency penalty would penalize new tokens based on their previous frequency in the text, reducing the likelihood that the model will repeat the same line verbatim. This value was also set to zero. For the instruction of the GPT model, we used prompts with zero-shot prompting [49] for the system role to have the LLM perform the desired tasks as a completion task. These settings applied equally to the use of the model’s vision capabilities and to text generation.

4. Content Retrieval from the Declarative Memory of ACT-R

We demonstrated the use of the GPT-4o model to retrieve specific recollections from the declarative memory of an ACT-R model using two examples. In the first example, we assumed an utterance of a human in a dialog with a social robot, where eventually the robot is supposed to search ‘its’ (ACT-R) memory for matching recollections based on this utterance. In the second example, the goal was to assign the visual impression of the robot to existing recollections. In both cases, we wanted to retrieve a suitable recollection—if available—and have the LLM use its content through prompt augmentation.

4.1. Example Application 1: Recollection Based on Text

The assumed scenario for this example application was as follows: a humanoid social robot is standing at a train station on one of the tracks and offers its services to travelers looking for information. Understanding the content of human questions—especially in relation to certain scenarios defined via the system prompt—and answering them with the help of an LLM generally works well. However, the LLM probably does not know all the details of the current scenario and has no individual recollection of recent encounters between humans and the robot and therefore has no experience of the particular situation. It is therefore advantageous if such factual knowledge can be provided to the LLM for questions about specific local conditions, for example, so that it can be incorporated into an answer that is otherwise perhaps freely formulated by the language model.
The LLM was instructed via system prompts to understand the basic setting and everything we needed it to output. A special feature of this application was the instruction to first create a summary of the question in the form of three keywords or phrases consisting of a few words, instead of a complete answer, and put them in square brackets for better programmatic handling. These keywords or phrases were passed as chunks to the cognitive model to search for a matching recollection. We stored memory chunks, including the suitable fact phrases for this test application, as a priori knowledge in the declarative memory. For each API call that was used to generate a response to the human and not keywords or phrases, the entire dialog part was transferred to the GPT model without the previously generated keywords or phrases so that the model could ‘remember’ what was previously said in the dialog and refer to it as it progressed. The dialog text returned by the API was forwarded to the robot’s voice and tablet output.

4.1.1. Prompting the LLM

The system prompt for the LLM consisted of explanations of the scenario at hand, in general, and the instruction to output three keywords or short phrases as chunks that adequately represent the essential content of the human’s question, separated by commas in square brackets; for example, for the question ‘Where can I find platform 1?’, corresponding keywords could be [Platform 1, Location, Direction Assistance].
For the scenario, we used a system prompt like ‘You are a robot, your name is Pepper. You offer help and information to travelers in a train station by asking them where they want to go and how you can help’.
In the case that a chunk generation was required, we augmented this system prompt with ‘You do not answer the question, but always create three keywords in square brackets that describe the traveler’s problem’.
Furthermore, in the situation where the ACT-R model could not find any suitable recollection in the declarative memory, the system prompt was augmented with ‘You tell the traveler that you do not know the answer to the current question and do not make up an answer under any circumstances’. This should prevent the language model from inventing a seemingly suitable answer without a factual basis.

4.1.2. Results

If a matching chunk was found in the declarative memory, the fact phrase contained in this chunk was transferred back to the robot application and used to augment the LLM’s system prompt, e.g., with ‘Platform 1 is the neighboring platform right here’, as shown in Figure 3. This fact phrase was used by GPT-4o to respond with, for example, ‘Platform 1 is right here, neighboring this information desk. You can easily access it by walking straight ahead’.
The LLM therefore used the fact correctly, albeit in its own wording and with some embellishment from the scenario conditions. Details of a dialog run can be seen in Figure 4, showing an output in the graphical tool of the QiSDK robot emulator [50]. The output of the keywords in square brackets was only for testing purposes and would not be spoken by the robot in a production application. Part of a corresponding trace output of the running ACT-R model is shown in Figure 5.

4.2. Example Application 2: Recollection Based on Vision

No specific scenario was assumed for this example. The system’s ability to retrieve recollections from declarative memory by recognizing visual impressions from the robot’s perspective should be demonstrated. The procedure was basically similar to that in example 1. Again, the LLM was instructed to create a description of the content in the form of three keywords, but this time it was about the content of an image taken with the robot’s camera equipment. The robot’s front camera took a picture every few seconds to get an actual impression of what the robot was seeing. To prepare a response to a human utterance based on visual impressions, the robot application sent the currently captured image to the GPT-4o model as a completion task, using the type ‘image_url’ instead of ‘text’.
The three image-related keywords from the model’s response were transmitted as chunks to the cognitive model to search for comparable chunks in declarative memory. Here, too, we have previously stored suitable memory chunks for our test scenario in the declarative memory. In the event of a match, the content of the ‘phrase’ chunk slot was transferred to the robot application for use in prompt augmentation.
In our test applications, we focused on the usage of recollections for the prompt augmentation of a GPT model. However, the reverse was also possible, using this system to write memory chunks and thus visual impressions, which are provided with keywords, into the declarative memory in order to retrieve them later in the outlined manner. Below we give an example of a system prompt that not only generates the three keywords, but also provides a ‘phrase’ to describe the image content.

4.2.1. Prompting the LLM

For chunk generation, including a phrase that describes what is contained in this image and is to be stored in the ‘phrase’ slot, we provided a system prompt like the following ‘I am robot, my name is Pepper. I do not answer but create three keywords followed by a sentence with a short description describing what is in this image. The three keywords are enclosed by square brackets. The short sentence is enclosed by round brackets, for example [person, indoor, computer] (A sentence with a short description describing what is in this image)’.
For augmenting the prompt with a recollection, we used, for example, ‘I am robot, my name is Pepper. I can see what is described in the following:’ followed by the phrase recalled from declarative memory. In this way, the LLM could receive and process information about visual recollections stored in the past if similar sensory impressions, and thus, similar keywords are currently present.

4.2.2. Results

Figure 6 shows the generation of individual chunks from the keywords ‘keyboard’, ‘indoor’, and ‘christmas tree’, and it presents in a frame the phrase ‘A person is sitting indoors in front of a christmas tree appearing to engage in conversation or video call’, supplied by the LLM in the running ACT-R model. These chunks were transferred to the declarative memory of the cognitive model, not only separately, but also as a composite chunk consisting of keywords and phrase. To be able to save keywords and phrases as chunks, spaces had to be replaced by a ‘-’. The procedure was reversed when the phrase was used for prompt augmentation. A retrieval of a recollection by the ACT-R model is shown in Figure 7 in framed areas. The model searches the stored memory chunks using the supplied keywords to find a matching recollection, the content of which is passed from the ‘phrase’ slot to the robot application for use in prompt augmentation.

5. Discussion

On the one hand, cognitive models can be used within the framework of their cognitive architectures to anticipate human behavior and thus make human cognitive processes more understandable—even for a robot. On the other hand, cognitive processes can use declarative and procedural memory to produce decisions which, when used to control the actions of a robot, allow the robot to act in a more humane way. Such processes can therefore be used to make human actions understandable for the machine, but also to make the robot’s actions more comprehensible for the human in HRI. Kambhampati introduced the term human-aware AI to refer to AI systems that are explainable to humans in the loop and capable of working synergistically with humans [51].
The possibility of using cognitive architectures to unlock human-like judgment and decision-making capabilities such as instance-based learning and intuitive decision making for social robots can be an opportunity for greater acceptance and trust through more common ground and similarities in the way we—humans and robots—think. With the help of LLMs and VLMs and the corresponding sensors, robots can access and interpret the same information that is available to humans. The use of declarative and procedural memory in cognitive models makes it possible to process the data provided by robot sensors and language models in a human-centered way. We only outline the possibilities here with our example applications for storing visual impressions or facts together with keywords and using them for the prompt augmentation of an LLM. In general, there are far more options available for a programmatic implementation of cognitive models with ACT-R or similar frameworks in combination with social robots. For example, one could consider using descriptions generated by a VLM of what the robot is currently seeing to match visual perceptions between humans and robots via memory functions of a cognitive model. Other examples include situation awareness systems and cognitive digital twins [52,53]. Harnessing the memory and decision-making capabilities of cognitive models can even be a step towards Continual Learning (CL) that can adapt to the ever-changing needs, preferences, and environments of users [54].
The proposed system and procedure is not without shortcomings and open problems. It is well known that LLMs and VLMs are sometimes prone to hallucinations. One possible way to reduce the risk of such hallucinations and the reproduction of made-up statements by an LLM, such as OpenAI’s GPT, would be to provide the language model with additional information tailored to individual scenarios via the system prompt, as we suggest in our application example based on remembered facts. However, even such a method of constraining the LLM does not offer absolute certainty for the exclusion of hallucinations.
To date, we have only tested the method with a few, mostly predefined memory contents, but we consider it scalable with regard to more complex cognitive models and the use of concepts such as forgetting, learning new facts, and reinforcement of recollections, or further possibilities made possible by the cognitive architecture of ACT-R.
The system works in real time and can be used for fluent dialogs between humans and robots. However, the time required to retrieve relevant recollections could play a limiting role, depending on the amount of memory chunks available. In addition, for each interaction with the human user, our system requires two consecutive API calls to the GPT model, both of which have a certain latency. Together, these time aspects may cause a noticeable and unnatural delay in interaction, even if there was no significant difference to conventional use with just one call to the GPT API in our tests.
Furthermore, it also seems to be a disadvantage and increases the complexity that the implementation of the cognitive architecture in our test applications runs as a standalone version on an extra computer and not on the robot itself. However, this problem could be solved by using other robot models and thus other ways of implementing the cognitive architecture, such as pyactr, as an ACT-R implementation for the Python-based robot Navel [55,56].

6. Conclusions

We propose a design and development approach for a combined system consisting of an ACT-R cognitive model and a humanoid social robot to endow the robot with human-like memory capabilities. Recollections from the declarative memory of the ACT-R model could be retrieved using real-world data obtained by the robot via an LLM or VLM. The procedural memory, which consists of the productions of the cognitive model, was used to retrieve these recollections and return them to the robot as instructions for action. In addition, real-world data captured by the robot could be stored as memory chunks in the cognitive model’s declarative memory.
In an example application, such a system was used to improve the correctness and accuracy of GPT-4o’s reasoning capabilities by using recollections for prompt augmentation. Another example application delivered keyword labeled visual impressions of the robot to the declarative memory or retrieved recollections based on these impressions. We tested this system with the social robot Pepper. In principle, however, this method can also be used independently of HRI scenarios.
Regarding the explainability aspects, our system provides approaches for a possible constraint of LLMs to generate robot utterances in HRI by comprehensible memory contents. The type of possible connection or integration of the cognitive model depends on the robot’s operating system and on whether or which ACT-R implementations or implementations of a comparable cognitive architecture are available for it. For example, there are direct implementations for Python, which means that the standalone version of ACT-R on an external computer and thus also the TCP/IP connection between the cognitive model and the robot application could be avoided.
The use of a cognitive architecture such as ACT-R enables the inclusion and investigation of further cognitive principles and processes in interaction with a social robot and LLMs independent of memory skills and declarative memory retrieval. Future work should consider these aspects in more detail. For the future, we are also planning studies to investigate metrics such as latencies in the use of the LLM and the cognitive model, as well as the precision of the associated memories and their use for prompt augmentation in more detail. An ablation study would also be helpful to investigate the added value and effectiveness of our approach more closely. Further experiments are needed to optimize the ACT-R models and system prompts for the LLM, as well as ongoing evaluation in studies with different people and different robot systems for various tasks. We are confident that this will open up a wide range of possibilities for future research into how cognitive architectures and their models can add a human touch to a social robot in HRI.

Author Contributions

Conceptualization, T.S. and N.R.; methodology, T.S. and N.R.; writing—original draft preparation, T.S.; writing—review and editing, T.S. and N.R.; supervision, N.R. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The raw data supporting the conclusions of this article will be made available by the authors upon request.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
ACT-RAdaptive Control of Thought-Rational
APIApplication Programming Interface
AIArtificial Intelligence
ACoArtificial Cognition
BNBayesian Network
CLContinual Learning
CoALACognitive Architectures for Language Agents
DLDeep Learning
GPTGenerative Pretrained Transformer
HRIHuman-Robot Interaction
IBLInstance-Based Learning Theory
LLMLarge Language Model
LMMLarge Multimodal Models
LTMLong-Term Memory
RAGRetrieval-Augmented Generation
RLReinforcement Learning
TCP/IPTransmission Control Protocol/Internet Protocol
ToMTheory of Mind
VLMVision-Language Model

References

  1. Bisaz, R.; Travaglia, A.; Alberini, C. The Neurobiological Bases of Memory Formation: From Physiological Conditions to Psychopathology. Psychopathology 2014, 47, 347–356. [Google Scholar] [CrossRef] [PubMed]
  2. Zhang, J. How memories are stored in the brain: The declarative memory model. arXiv 2024, arXiv:2403.17985. [Google Scholar]
  3. Myers, N.E.; Stokes, M.G.; Nobre, A.C. Prioritizing Information during Working Memory: Beyond Sustained Internal Attention. Trends Cogn. Sci. 2017, 21, 449–461. [Google Scholar] [CrossRef]
  4. Braun, E.K.; Wimmer, G.E.; Shohamy, D. Retroactive and graded prioritization of memory by reward. Nat. Commun. 2018, 9, 4886. [Google Scholar] [CrossRef]
  5. Sandini, G.; Sciutti, A.; Morasso, P. Artificial cognition vs. artificial intelligence for next-generation autonomous robotic agents. Front. Comput. Neurosci. 2024, 18, 1349408. [Google Scholar] [CrossRef]
  6. Dubey, S.; Ghosh, R.; Dubey, M.; Chatterjee, S.; Das, S.; Benito-León, J. Redefining Cognitive Domains in the Era of ChatGPT: A Comprehensive Analysis of Artificial Intelligence’s Influence and Future Implications. Med. Res. Arch. 2024, 12, 5383. [Google Scholar] [CrossRef]
  7. Janik, R.A. Aspects of human memory and Large Language Models. arXiv 2024, arXiv:2311.03839. [Google Scholar]
  8. Jeong, H.; Lee, H.; Kim, C.; Shin, S. A Survey of Robot Intelligence with Large Language Models. Appl. Sci. 2024, 14, 8868. [Google Scholar] [CrossRef]
  9. Ghosh, A.; Acharya, A.; Saha, S.; Jain, V.; Chadha, A. Exploring the Frontier of Vision-Language Models: A Survey of Current Methodologies and Future Directions. arXiv 2024, arXiv:2404.07214. [Google Scholar]
  10. Kawaharazuka, K.; Obinata, Y.; Kanazawa, N.; Okada, K.; Inaba, M. Robotic Applications of Pre-Trained Vision-Language Models to Various Recognition Behaviors. In Proceedings of the 2023 IEEE-RAS 22nd International Conference on Humanoid Robots (Humanoids), Austin, TX, USA, 12–14 December 2023; pp. 1–8. [Google Scholar] [CrossRef]
  11. Tan, R.; Yang, H.; Jiao, S.; Shan, L.; Tao, C.; Jiao, R. Smart Robot Manipulation using GPT-4o Vision. In Proceedings of the 7th European Industrial Engineering and Operations Management Conference, Augsburg, Germany, 16–18 July 2024; IEOM Society International: Southfield, MI, USA, 2024. [Google Scholar] [CrossRef]
  12. Kurup, U.; Lebiere, C. What can cognitive architectures do for robotics? Biol. Inspired Cogn. Archit. 2012, 2, 88–99. [Google Scholar] [CrossRef]
  13. Anderson, J.R.; Bothell, D.; Byrne, M.D.; Douglass, S.; Lebiere, C.; Qin, Y. An integrated theory of the mind. Psychol. Rev. 2004, 111, 1036–1060. [Google Scholar] [CrossRef] [PubMed]
  14. Thomson, R.; Lebiere, C.; Anderson, J.R.; Staszewski, J. A general instance-based learning framework for studying intuitive decision-making in a cognitive architecture. J. Appl. Res. Mem. Cogn. 2015, 4, 180–190. [Google Scholar] [CrossRef]
  15. Gonzalez, C.; Dutt, V.; Lebiere, C. Validating instance-based learning mechanisms outside of ACT-R. J. Comput. Sci. 2013, 4, 262–268. [Google Scholar] [CrossRef]
  16. Gonzalez, C.; Lerch, J.F.; Lebiere, C. Instance-based learning in dynamic decision making. Cogn. Sci. 2003, 27, 591–635. [Google Scholar] [CrossRef]
  17. Werk, A.; Scholz, S.; Sievers, T.; Russwinkel, N. How to Provide a Dynamic Cognitive Person Model of a Human Collaboration Partner to a Pepper Robot. In Proceedings of the ICCM, Tilburg, The Netherlands, 19–22 July 2024. [Google Scholar]
  18. Kahneman, D. Thinking, Fast and Slow; Farrar, Straus and Giroux: New York, NY, USA, 2011. [Google Scholar]
  19. Whitehill, J. Understanding ACT-R—An Outsider’s Perspective. arXiv 2013, arXiv:1306.0125. [Google Scholar]
  20. Quam, C.; Wang, A.; Maddox, W.T.; Golisch, K.; Lotto, A. Procedural-Memory, Working-Memory, and Declarative-Memory Skills Are Each Associated With Dimensional Integration in Sound-Category Learning. Front. Psychol. 2018, 9, 1828. [Google Scholar] [CrossRef]
  21. Eichenbaum, H. Hippocampus: Cognitive Processes and Neural Representations that Underlie Declarative Memory. Neuron 2004, 44, 109–120. [Google Scholar] [CrossRef]
  22. Hayne, H.; Boniface, J.; Barr, R. The Development of Declarative Memory in Human Infants: Age-Related Changes in Deferred Imitation. Behav. Neurosci. 2000, 114, 77–83. [Google Scholar] [CrossRef]
  23. OpenAI. The Most Powerful Platform for Building AI Products. 2025. Available online: https://openai.com/api/ (accessed on 1 February 2025).
  24. Carnegie Mellon University. ACT-R Software. 2024. Available online: http://act-r.psy.cmu.edu/software/ (accessed on 10 December 2024).
  25. Sievers, T.; Russwinkel, N. How to Use a Cognitive Architecture for a Dynamic Person Model with a Social Robot in Human Collaboration. CEUR Workshop Proceedings. 2024. Available online: https://ceur-ws.org/Vol-3794/paper2.pdf (accessed on 23 October 2024).
  26. Lykov, A.; Litvinov, M.; Konenkov, M.; Prochii, R.; Burtsev, N.; Abdulkarim, A.A.; Bazhenov, A.; Berman, V.; Tsetserukou, D. CognitiveDog: Large Multimodal Model Based System to Translate Vision and Language into Action of Quadruped Robot; HRI ’24; Association for Computing Machinery: New York, NY, USA, 2024; pp. 712–716. ISBN 9798400703232. [Google Scholar] [CrossRef]
  27. Wang, J.; Shi, E.; Hu, H.; Ma, C.; Liu, Y.; Wang, X.; Yao, Y.; Liu, X.; Ge, B.; Zhang, S. Large language models for robotics: Opportunities, challenges, and perspectives. J. Autom. Intell. 2024, 4, 52–64. [Google Scholar] [CrossRef]
  28. Hu, Y.; Lin, F.; Zhang, T.; Yi, L.; Gao, Y. Look Before You Leap: Unveiling the Power of GPT-4V in Robotic Vision-Language Planning. arXiv 2023, arXiv:2311.17842. [Google Scholar]
  29. Wake, N.; Kanehira, A.; Sasabuchi, K.; Takamatsu, J.; Ikeuchi, K. GPT-4V(ision) for Robotics: Multimodal Task Planning from Human Demonstration. arXiv 2024, arXiv:2311.12015. [Google Scholar] [CrossRef]
  30. Yoshida, T.; Baba, S.; Masumori, A.; Ikegami, T. Minimal Self in Humanoid Robot “Alter3” Driven by Large Language Model. arXiv 2024, arXiv:2406.11420. [Google Scholar]
  31. Leidner, D. Toward Robotic Metacognition: Redefining Self-Awareness in an Era of Vision-Language Models. In Proceedings of the 40th Anniversary of the IEEE Conference on Robotics and Automation, Rotterdam, The Netherlands, 23–26 September 2024. [Google Scholar]
  32. Niu, Q.; Liu, J.; Bi, Z.; Feng, P.; Peng, B.; Chen, K.; Li, M.; Yan, L.K.; Zhang, Y.; Yin, C.H.; et al. Large Language Models and Cognitive Science: A Comprehensive Review of Similarities, Differences, and Challenges. arXiv 2013, arXiv:1306.0125. [Google Scholar]
  33. Wu, S.; Oltramari, A.; Francis, J.; Giles, C.L.; Ritter, F.E. Cognitive LLMs: Towards Integrating Cognitive Architectures and Large Language Models for Manufacturing Decision-making. arXiv 2024, arXiv:2408.09176. [Google Scholar]
  34. González-Santamarta, M.Á.; Rodríguez-Lera, F.J.; Guerrero-Higueras, Á.M.; Matellán-Olivera, V. Integration of Large Language Models within Cognitive Architectures for Autonomous Robots. arXiv 2023, arXiv:2309.14945. [Google Scholar]
  35. Berto, L.; Tanevska, A.; Cirne, A.; Costa, P.; Simões, A.; Gudwin, R.; Rea, F.; Colombini, E.; Sciutti, A. Curiosity and Affect-Driven Cognitive Architecture for HRI. IEEE Trans. Affect. Comput. 2025, 1–18. [Google Scholar] [CrossRef]
  36. Hellou, M.; Vinanzi, S.; Cangelosi, A. A Review of Theory of Mind and Robotics: Mind Reading in Human-Robot Interaction for Proactive Social Robots. In Human-Friendly Robotics 2024; Paolillo, A., Giusti, A., Abbate, G., Eds.; Springer: Cham, Switzerland, 2025; pp. 197–211. [Google Scholar]
  37. Sievers, T.; Russwinkel, N.; Möller, R. Grounding a Social Robot’s Understanding of Words with Associations in a Cognitive Architecture. In Proceedings of the 17th International Conference on Agents and Artificial Intelligence, Porto, Portugal, 23–25 February 2025; pp. 406–410. [Google Scholar] [CrossRef]
  38. Knowles, K.; Witbrock, M.; Dobbie, G.; Yogarajan, V. A Proposal for a Language Model Based Cognitive Architecture. Proc. AAAI Symp. Ser. 2024, 2, 295–301. [Google Scholar] [CrossRef]
  39. Leivada, E.; Marcus, G.; Günther, F.; Murphy, E. A Sentence is Worth a Thousand Pictures: Can Large Language Models Understand Human Language and the World behind Words? arXiv 2023, arXiv:2308.00109. [Google Scholar]
  40. Sun, R. Can A Cognitive Architecture Fundamentally Enhance LLMs? Or Vice Versa? arXiv 2024, arXiv:2401.10444. [Google Scholar]
  41. He, Z.; Lin, W.; Zheng, H.; Zhang, F.; Jones, M.W.; Aitchison, L.; Xu, X.; Liu, M.; Kristensson, P.O.; Shen, J. Human-inspired Perspectives: A Survey on AI Long-term Memory. arXiv 2024, arXiv:2411.00489. [Google Scholar]
  42. Jiang, X.; Li, F.; Zhao, H.; Wang, J.; Shao, J.; Xu, S.; Zhang, S.; Chen, W.; Tang, X.; Chen, Y.; et al. Long Term Memory: The Foundation of AI Self-Evolution. arXiv 2024, arXiv:2410.15665. [Google Scholar]
  43. Hou, Y.; Tamoto, H.; Miyashita, H. “My agent understands me better”: Integrating Dynamic Human-like Memory Recall and Consolidation in LLM-Based Agents. In Proceedings of the Extended Abstracts of the CHI Conference on Human Factors in Computing Systems, CHI EA ’24, Honolulu, HI, USA, 11–16 May 2024; Association for Computing Machinery: New York, NY, USA, 2024. [Google Scholar] [CrossRef]
  44. Sumers, T.R.; Yao, S.; Narasimhan, K.; Griffiths, T.L. Cognitive Architectures for Language Agents. arXiv 2024, arXiv:2309.02427. [Google Scholar]
  45. Gao, Y.; Xiong, Y.; Gao, X.; Jia, K.; Pan, J.; Bi, Y.; Dai, Y.; Sun, J.; Wang, M.; Wang, H. Retrieval-Augmented Generation for Large Language Models: A Survey. arXiv 2024, arXiv:2312.10997. [Google Scholar]
  46. Sanmartin, D. KG-RAG: Bridging the Gap Between Knowledge and Creativity. arXiv 2024, arXiv:2405.12035. [Google Scholar]
  47. Trafton, G.; Hiatt, L.; Harrison, A.; Tanborello, F.; Khemlani, S.; Schultz, A. ACT-R/E: An embodied cognitive architecture for human–robot interaction. J. Hum.-Robot Interact. 2013, 2, 30–55. [Google Scholar] [CrossRef]
  48. Aldebaran, United Robotics Group and Softbank Robotics. Pepper. 2025. Available online: https://www.aldebaran.com/en/pepper (accessed on 3 January 2025).
  49. Brown, T.; Mann, B.; Ryder, N.; Subbiah, M.; Kaplan, J.; Dhariwal, P.; Neelakantan, A.; Shyam, P.; Sastry, G.; Askell, A.; et al. Language Models are Few-Shot Learners. arXiv 2020, arXiv:2005.14165. [Google Scholar] [CrossRef]
  50. Aldebaran, United Robotics Group and Softbank Robotics. Pepper SDK for Android. 2025. Available online: https://qisdk.softbankrobotics.com/sdk/doc/pepper-sdk/index.html (accessed on 21 January 2025).
  51. Sreedharan, S.; Kulkarni, A.; Kambhampati, S. Explainable Human-AI Interaction: A Planning Perspective. arXiv 2024, arXiv:2405.15804. [Google Scholar]
  52. Russwinkel, N. A Cognitive Digital Twin for Intention Anticipation in Human-Aware AI. In Intelligent Autonomous Systems 18; Lee, S.G., An, J., Chong, N.Y., Strand, M., Kim, J.H., Eds.; Springer: Cham, Switzerland, 2024; pp. 637–646. [Google Scholar]
  53. Hao, C.; Russwinkel, N.; Haeufle, D.; Beckerle, P. Conceptualizing the Integration of Cognitive and Physical Models to Enable Continuous Human-Robot Interaction. 2025; forthcoming. [Google Scholar]
  54. Ayub, A.; De Francesco, Z.; Mehta, J.; Yaakoub Agha, K.; Holthaus, P.; Nehaniv, C.L.; Dautenhahn, K. A Human-Centered View of Continual Learning: Understanding Interactions, Teaching Patterns, and Perceptions of Human Users Toward a Continual Learning Robot in Repeated Interactions. J. Hum.-Robot Interact. 2024, 13, 1–39. [Google Scholar] [CrossRef]
  55. Jakdot. Pyactr. 2025. Available online: https://github.com/jakdot/pyactr (accessed on 1 April 2025).
  56. Navel Robotics GmbH. Navel. 2025. Available online: https://navelrobotics.com (accessed on 1 April 2025).
Figure 1. Connection between ACT-R/dispatcher and the robot application.
Figure 1. Connection between ACT-R/dispatcher and the robot application.
Applsci 15 05778 g001
Figure 2. Pepper showing a picture of what it sees in dialog with a human.
Figure 2. Pepper showing a picture of what it sees in dialog with a human.
Applsci 15 05778 g002
Figure 3. Transfer of keywords to ACT-R and return of retrieved memory facts to the LLM for prompt augmentation.
Figure 3. Transfer of keywords to ACT-R and return of retrieved memory facts to the LLM for prompt augmentation.
Applsci 15 05778 g003
Figure 4. Dialog output in the graphical tool of the QiSDK robot emulator.
Figure 4. Dialog output in the graphical tool of the QiSDK robot emulator.
Applsci 15 05778 g004
Figure 5. ACT-R trace output of the running model.
Figure 5. ACT-R trace output of the running model.
Applsci 15 05778 g005
Figure 6. ACT-R model creating chunks for keywords and phrase.
Figure 6. ACT-R model creating chunks for keywords and phrase.
Applsci 15 05778 g006
Figure 7. ACT-R model retrieving chunks with keywords and phrases from declarative memory.
Figure 7. ACT-R model retrieving chunks with keywords and phrases from declarative memory.
Applsci 15 05778 g007
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Sievers, T.; Russwinkel, N. Retrieving Memory Content from a Cognitive Architecture by Impressions from Language Models for Use in a Social Robot. Appl. Sci. 2025, 15, 5778. https://doi.org/10.3390/app15105778

AMA Style

Sievers T, Russwinkel N. Retrieving Memory Content from a Cognitive Architecture by Impressions from Language Models for Use in a Social Robot. Applied Sciences. 2025; 15(10):5778. https://doi.org/10.3390/app15105778

Chicago/Turabian Style

Sievers, Thomas, and Nele Russwinkel. 2025. "Retrieving Memory Content from a Cognitive Architecture by Impressions from Language Models for Use in a Social Robot" Applied Sciences 15, no. 10: 5778. https://doi.org/10.3390/app15105778

APA Style

Sievers, T., & Russwinkel, N. (2025). Retrieving Memory Content from a Cognitive Architecture by Impressions from Language Models for Use in a Social Robot. Applied Sciences, 15(10), 5778. https://doi.org/10.3390/app15105778

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop