Next Article in Journal
Comparative Analysis of the Spatiotemporal Evolution Patterns of Acoustic Emission Source Localization Under True Triaxial Loading and Loading-Unloading Conditions in Sandstone
Previous Article in Journal
Moving-Target Tracking in Airport Airside Operations Using AIMM-STUKF
Previous Article in Special Issue
Robot-Assisted Mirror Rehabilitation for Post-Stroke Upper Limbs: A Personalized Control Strategy
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Cross-Modality Alignment Perception and Multi-Head Self-Attention Mechanism for Vision-Language-Action of Humanoid Robot

1
Shanghai Key Laboratory of Intelligent Manufacturing and Robotics, School of Mechatronic Engineering and Automation, Shanghai University, Shanghai 200444, China
2
Zhejiang Key Laboratory of Robotics and Intelligent Manufacturing Equipment Technology, Ningbo Institute of Materials Technology & Engineering, Chinese Academy of Sciences, Ningbo 315201, China
*
Author to whom correspondence should be addressed.
Sensors 2026, 26(1), 165; https://doi.org/10.3390/s26010165 (registering DOI)
Submission received: 24 October 2025 / Revised: 7 December 2025 / Accepted: 8 December 2025 / Published: 26 December 2025

Highlights

What are the main findings?
  • A model of memory-gated filtering attention was proposed, which improved multi-head self-attention mechanism.
  • A cross-modal alignment perception during training was designed, which combined with a few-shot data collection strategy of key steps.
What are the implications of the main findings?
  • The proposed multi-head self-attention mechanism could reduce the video memory occupation by 72% and improve the training speed from 1.35 s to 0.129 s per batch.
  • The proposed Vision-Language-Action of a humanoid robot significantly improved the task success rate and alleviated the robot arm jitter problem.

Abstract

For a humanoid robot, it is difficult to predict a motion trajectory through end-to-end imitation learning when performing complex operations and multi-step processes, leading to jittering in the robot arm. To alleviate this problem and reduce the computational complexity of the self-attention module in Vision-Language-Action (VLA) operations, we proposed a memory-gated filtering attention model that improved the multi-head self-attention mechanism. Then, we designed a cross-modal alignment perception during training, combined with a few-shot data-collection strategy for key steps. The experimental results showed that the proposed scheme significantly improved the task success rate and alleviated the robot arm jitter problem, while reducing video memory usage by 72% and improving training speed from 1.35 s to 0.129 s per batch. This maintained higher action accuracy and robustness in the humanoid robot.

1. Introduction

Confronting complex tasks in industrial scenarios is often challenging for embodied intelligent robots, as they require accuracy, generalization across tasks, careful coordination of contact forces, and closed-loop visual feedback [1]. To perform these tasks, the embodied intelligent robot needs not only to master multiple actions but also to execute the appropriate task based on camera images and language instructions, with high accuracy [2]. The manipulation tasks of embodied intelligent robots often involve multiple challenges, including multimodal perception, long-term temporal dependencies, uncertainty in physical interactions, and adaptability to dynamic environments. These features place higher demands on existing data-driven control strategies [3,4].
For robots processing complex long-step tasks, imitation learning (IL) has become one of the important technical approaches by advancing robot operation to actual deployment, as it can directly learn the perception-action mapping from human demonstrations and reduce the costs of complex symbolic modeling and online exploration [5]. Especially in high-dimensional, contact-rich manipulation scenarios, such as insertion, wearing, assembly, and flexible object manipulation, imitation learning is often more feasible and data efficient than reinforcement learning (RL) [6].
However, imitation learning has many limitations. For instance, previous methods [7,8] have demonstrated strong imitation learning performance in open scenarios; however, they still face significant challenges in long-term, multi-step tasks. In particular, one-step BC is prone to compounding errors in multi-step tasks; small prediction biases are gradually amplified over long sequences, leading to failure in the final operation [9,10,11,12]. To address this, recent studies have adopted structured prediction or trajectory-generation mechanisms; for example, combining visual perception with trajectory generation has been explored in robotic arm tasks such as grasping and assembly [5]. Furthermore, experiments with low-cost, low-precision manipulators in unstructured environments show that although imitation learning can achieve initial control with few demonstrations, its robustness decreases significantly as task complexity and continuous control requirements increase [13,14,15].
In recent years, inspired by the breakthrough results of Large Language Models (LLMs) in natural language understanding and reasoning, researchers have gradually introduced language model architectures into the field of embodied intelligence. Attempts have been made to integrate perception, planning, and control capabilities within a unified model framework to improve robots’ semantic understanding and autonomous decision-making in complex, dynamic environments [16,17,18,19,20]. Related work includes both a hybrid architecture (language-as-policy), where an LLM is used for high-level task decomposition and specialized control modules perform low-level actions, and a more tightly coupled closed-loop sense-reasoning-control pipeline. Multimodal observations, such as vision and touch, are continuously fed back to the language module to support online plan revision and action generation [16,17]. In addition, several reviews and empirical studies have shown that combining LLM or multimodal LLM (VLM/MLLM) with world models, retrieval augmentation mechanisms, or security constraint modules can improve generalization and robustness in long-duration tasks, few-shot or zero-shot scenarios, but real-time performance, security, and physical constraint alignment are still major challenges [18,19,20].
In summary, the contributions of this paper to the abovementioned work were as follows:
(1)
For the model algorithm, we improved the multi-head self-attention mechanism [21] and proposed memory-gated filtering attention, which not only reduced the video memory occupation but also significantly accelerated the training speed with improved performance and reduced the algorithm complexity.
(2)
In terms of datasets, the current end-to-end collection strategy collected all trajectories of the manipulator from the starting point to the target point [22], and there were many noises and redundancies in the intermediate trajectory set. We designed and collected data on key steps, such as the task start point and goal point of the mechanical claw reaching the manipulated object, which can suppress the mechanical arm’s jitter during model inference and enable anthropomorphic operation.
(3)
In the training strategy, in order to deal with the sudden addition of a new environmental observation [11] after the robot performs an action each time, which could cause the robot to move unsteadily, we proposed cross-modal alignment sensing. For each key observation frame, the model not only learns to predict the single action strictly corresponding to it but also aligns all possible subsequent remaining action sequences.

2. Vision-Language-Action (VLA) for Improved Multi-Head Self-Attention Mechanism

2.1. Memory-Gated Filtering Attention

In the robot operation task, the VLA operation could receive observation information from the environment and a natural-language instruction as input. After multimodal encoding and fusion, the corresponding action sequence was generated for robot execution, the overall process shown in Figure 1.
To train the memory-gated filtering attention Vision-Language-Action (VLA) model based on a multi-head attention mechanism, we designed a unified encoding process for multi-modal observation data, ensuring that information from the visual, language, and action channels were fully aligned and interacted in the high-dimensional semantic space. The transformer architecture was served as the core processing framework for individual modalities [23], illustrated in Figure 2.
Among them, self-attention was the core mechanism of the transformer model. However, the self-attention algorithm had two particularly obvious shortcomings. The first one was the problem of high computational complexity caused by parallel computing. Its time complexity was O ( N 2 × D k ).
In each head, the query matrix q i was first dotted with a learnable parameter matrix W f , and the sigmoid function was used for activation to generate the gating filter matrix s i g m o d q i W f . Subsequently, the row-wise accumulated k i matrix c o s u m k i undergoes element-wise multiplication (Hadamard product) with this gating matrix for filtering. Finally, the output matrix s i g m o d q i W f c o s u m k i was concatenated with the query matrix   q i , and the combined result was fused through a linear layer W o to obtain the Mem i   matrix for each head. The outputs of all heads were then concatenated together to yield MemAttn , as shown in the following equation:
  Mem i = [ q i   |   s i g m o d q i W f c o s u m k i ] W o
        MemAttn = cat Mem 1 , Mem 2 , , Mem h
Among them, W f R n × d h , W o R 2 d h × d h , for the training parameters, c o s u m k i = k 1 + + k i ,     MemAttn R n × d , q i , k i R n × d h , n for sequence length d as the feature dimension, h is the number of heads.

2.2. Multi-Modal Fusion of Vision-Language-Action

First, we could take an observation, each of which consists of an RGB image with a resolution of 640 × 480 and a depth map perfectly aligned with its pixels. In particular, depth information was normalized during preprocessing to ensure stable depth value distributions across different scenes. The natural language instructions corresponding to the task were used as conditional information to participate in subsequent cross-modal interactions. In the input construction stage, the RGB image with size 3 × 640 × 480 and the depth map with size 1 × 640 × 480 were concatenated along the channel dimension to obtain a complete 4 × 640 × 480 RGBD observation tensor.
To further introduce local spatial perception and reduce the input dimension, we adopted the same Patch Embedding strategy as Vision Transformer (ViT) [24]. This was performed as follows:
(1)
Each observation data was divided into multiple image patches of fixed size, and the size of each patch was 4 × m × m, containing data of four channels of [R, G, B, Depth].
(2)
Each patch was flattened into a one-dimensional vector in the spatial dimension, with a size of 1   ×   4   m 2 .
(3)
Therefore, a total of 640   ×   480 m   ×   m patches can be obtained for the observed data with a resolution of 640 × 480.
After patch flattening, the representation of the whole frame of observation data becomes:   Vision _ vit 640   ×   480 m   ×   m , 4   m 2 .
Then, the representation was fed into the Multi-Layer Perceptron (MLP) linear layer, and dimension transformation and feature projection were performed to map each patch into a unified semantic space, yielding the final visual input tensor. Where d m o d e l represents the representation dimension defined in the model.
Vision = L i n e a r Vision _ vit
Vision = v 1 , v 2 , , v d m o d e l , shape = 640 × 480 m × m , d m o d e l
After the observation data underwent Patch Embedding and MLP projection, the resulting visual representation was fed into a visual encoder module based on a transformer architecture for feature interaction with spatial context. In this stage, the visual sequence was interacted with multi-layer memory-gated filtering attention to preserve the contextual semantic information across local and global perception. Considering the design characteristics of memory-gated filtering attention, only the feature representation of the last time step should be retained as a compact representation of the whole frame observation data in the output of the whole visual encoder, which is denoted as:
V i s i o n l a s t R 1 × d m o d e l
This representation aggregated the multi-scale information of all patches in the current frame as a high-dimensional semantic embedding of the environment state at that instant.
In the multi-modal fusion stage, the V i s i o n l a s t output from the visual encoder and the corresponding natural language instructions were jointly input into the language encoder. The language encoder also leverages the pre-trained parameters of the large language model to fully exploit the model’s knowledge transfer capabilities for natural language understanding and to enhance its adaptability across diverse task descriptions and scene contexts. Let a given instruction be l, which is passed through tokenizer and embedding to obtain a sequence of word vectors e = e 1 , e 2 , , e t .
For cross-modal alignment, the visual global feature V i s i o n l a s t was inserted at the top of the sequence and concatenated with the language feature sequence as follows:
  V i s i o n l a s t , e 1 , e 2 , , e t , shape = 1 + t , d model
Then, the concatenated multimodal sequences were sent to the language encoder, and through interactions within the multi-layer transformer block, the deep fusion and complementarity of vision-language information are further realized. Finally, only the output of the last time step after fusion was retained as a compact representation of the cross-modal global context semantics, denoted as follows:
                  V L l a s t R 1 × d m o d e l
After the previous steps of visual language instruction processing, the next step is action generation. For action, we have seven degrees of freedom for one arm and 14 degrees of freedom for two arms in total, and each action data is denoted as a t =   a 1 , a 2 , , a 14 . In an observation, we have k steps of action. Before processing, we need to project the action using an MLP linear layer to increase the dimensionality.
    a d = L i n e a r a t
a d i =   a d i 1 , a d i 2 , , a di model , a d R k × d m o d e l
Similarly, we concatenated the output of the last time step of the visual language to the sequence head in the action data to be sent to the action decoder for training, labeled as follows:
    V L l a s t , a d 1 , a d 2 , , a d k , s h a p e = 1 + k , d m o d e l
The fused visual-language action data   V L l a s t , a d 1 , a d 2 , , a d k   was input to the action-generation decoder. We also used a memory-gated filtering attention module based on a pre-trained large language model for sequence modeling and action generation. By introducing a memory enhancement mechanism, the module establishes stronger context dependence across different action time steps, thereby improving the consistency and stability of multi-step control signal generation in complex operation tasks.

3. Cross-Modal Alignment Awareness Strategy

3.1. Data Collection for Imitation Learning

In the data collection phase of a robot manipulation task, we can collect a motion trajectory that matches the natural language instruction l , the observed data I within   t N during task execution, and the corresponding action sequence a t [25,26]. The existing end-to-end acquisition strategies are generally used to record the complete action trajectory of the robot from the starting state to the completion of the task in a time interval. For example, for the instruction “let the robot open the drawer”, the data collection often covers the whole process from identifying the position of the drawer, approaching, grasping to the drawer being completely pulled open, that is, the collection of the continuous action sequence a t = a 0 , , a r [27,28] under the given language instruction l, when the time period t = t 0 , t r .
However, in practice, we found that such a long time window trajectory acquisition method often introduces more noise data [29,30], and may lead to the lack of continuity of the generated joint control commands in the model inference stage, which will lead to problems such as joint jitter and trajectory offset during the execution process [31,32]. For the above problem, for example, under the given language instruction l , in the action sequence a t = a 0 , , a i , , a r , the key actions are a i and a r ; then we only need to collect   a i and a r , and the collected action sequence a t becomes a t = a i , a r . The intermediate action sequences a 0 , , a i and a i , , a r , without model inference, only need to let the manipulator reach the key position by itself.
The manipulator will perform multiple action adjustments to the environment before grasping. The mechanical claw’s posture data at the turning point in each path adjustment and the posture information at the final grasp are the “key steps” we need to collect during the task data-collection stage. The figure below shows two simple examples of collecting action-sequence data while the robot grasps a cube, compared to the existing way of working. The following figure shows a simple example.
After grasping, the robot needs to perform further specific interactive behaviors, such as putting the object into the container, performing insertion operations, or completing dumping. This stage also faces many challenges, especially in tasks with limited operating space or high-precision docking; a single action is often insufficient to complete the entire interaction. Therefore, the robot needs to dynamically adjust the end-effector’s posture and motion path in response to environmental feedback during operation. The image below shows the robot dropping the squares into an open drawer, shown in Figure 3.
The proposed strategy could effectively reduce noise from redundant actions, ensure continuous controllability of joint states, and reduce the complexity of motion planning and generation, thereby improving the robustness and interpretability of robot operations driven by natural language commands.
The observation data I consists of multimodal sensor information, specifically including a frame of RGBD image data of size 640 × 480. Each frame of observation is captured by a single depth camera and contains depth maps that are fully aligned with the color image, ensuring consistent encoding of the scene’s color and geometric information in the same spatial reference frame I = c a t I m a g e , d e p t h .
For sampling observation data, it is necessary to balance these two effects. We adopted the observation update strategy every two key steps; that is, we can use one frame of observation data to guide action generation for two consecutive key steps, thereby significantly improving overall task completion rate while maintaining environmental adaptability and action consistency. It greatly reduces the jump phenomenon of the action sequence at the observation insertion point and makes the robot’s behavior more coherent and natural during the execution of complex tasks, shown in Figure 4.
Through this phased, event-driven observation data-collection method, the robot can plan actions appropriately while maintaining efficient perception of the environment state, thereby better realizing sequential operation and real-time adaptation under complex instructions.

3.2. Cross-Modal Alignment of Datasets

After collecting all the natural language instructions l , observations I , and action sequences a t we need to align them for training. Our alignment strategy is to assume that we have collected a set of observation   data   I i = I i 0 , , I i m and action trajectory data a i = a i 0 , , a i n under the current natural language instruction l i . I i observation data and corresponding relationship with the trajectory a i for I i 0 a i 0 , a i 1 , I i 1 a i 2 , a i 3 ,..., I i m a i n 1 , a i n , where the correspondence between n and m, n = 2m + 1. In order to further improve the consistency of action generation and the coherence of multi-step actions in the inference stage of the model, we no longer uses the data alignment strategy that only aligns the observed data with the corresponding action one by one in the existing work, because the sudden addition of a new environmental observation after each step may lead to the unstable motion of the robot. So, we design the observation-action pairing strategy for cross-stage action alignment.
Specifically, in the training phase, for each key observation frame, the proposed VLA model not only learns to predict the single action strictly corresponding to it but also aligns all possible subsequent remaining action sequences. For example, the first observation corresponds to the entire sequence of actions a i 0 , a i 1 , , a i n , the second observation corresponds to the rest of the sequence except for the first two steps a i 2 , a i 3 , , a i n , and so on. This method effectively enhances the model’s ability to grasp the global task structure during reasoning, helps to generate smoother and continuous control instructions, and avoids mutation or jitter in the action sequence.
To ensure consistent tensor dimensions between the input and output in batch processing, padding must be applied to action sequences with different lengths. However, unlike the padding mask commonly used in NLP tasks, simply ignoring the padding term in the loss function may lead to non-physical action predictions in the continuous robot control scenario. Therefore, in this paper, the “standby action” a p   after the task is completed is introduced as a semantically reasonable filling value for the action length when alignment is insufficient. Specifically, if the length of the valid action sequence corresponding to an observation frame is less than the maximum length, it will be filled with several a p at the end of the sequence, which is formally expressed as:
  I i t a i 2 t , a i 2 t + 1 , , a i n , a p , , a p
where a p represents the stable standby state of the robot after the task is completed, which can physically ensure the execution continuity and avoid unnecessary action jitter or unstable output. In this way, the alignment of the observation data and the motion trajectory is as follows:
I i 0 I i 1 I i 2 I i m = a i 0 a i 1 a i 2 a i 3 a i n a i 2 a i 3 a i 4 a i 5 a p a i 4 a i 5 a i 6 a i 7 a p a i n 1 a i n a p a p a p

4. Complex Task Design of Humanoid Robot

4.1. Complex Task

To systematically evaluate the generalization ability and execution accuracy of the proposed VLA model on complex, multi-stage dual-arm manipulation tasks, we designed and selected seven common and challenging robotic tasks that could cover multiple types of object manipulation, spatial coordination, and cross-modal information fusion. Each task required the cooperation of two manipulators and high demands on language understanding, spatial perception, and fine control. These tasks included placing objects, opening and closing drawers, grasping targets, liquid dumping, and container operation, shown in Figure 5, Figure 6, Figure 7, Figure 8, Figure 9, Figure 10 and Figure 11. The descriptions of task segmentations were as follows:
(1)
“Put the red square into the drawer.” The shelf has two layers; the red square is randomly placed on the first or second shelf. (subtask #1) The left arm opens the drawer (subtask #2), the right arm locates the red square and grabs it (subtask #3), the right mechanical claw places the square into the drawer (subtask #4), and the left arm closes the drawer.
(2)
“Pour the water from the bottle into the cup.” There is a bottle full of water in front of you, and an empty paper cup. (subtask #1) Right arm looks for and grabs the bottle with water in it (subtask #2), picks up the water bottle and approaches the paper cup (subtask #3), evenly pours the water from the cup into the paper cup (subtask #4), and right arm puts the cup back in place after pouring the water.
(3)
“Put a spoon into a cup,” with a spoon in front of an empty paper cup. Only one spoon can be placed in the cup (subtask #1). The left arm locates the spoon and grasps it (subtask #2). It stops over the mouth of the cup without any error (subtask #3), puts the spoon into the cup, and raises and draws back the arm (subtask #4).
(4)
“Wipe the table with paper towels.” There is a paper towel on the table with a random, irregular stain. When pumping the paper towel, both arms need to stay close together to prevent the paper towel from being lifted. (Subtask #1) The left arm looks for the tissue and holds it, while the right hand presses the paper bag (subtask #2). The left arm removes the tissue and moves to the stained area, and the right hand returns to its original position (subtask #3). Left arm repeatedly wipes the stained area until it is clean.
(5)
“Put garbage bags and drugs into different drawers”. There are drugs on the first shelf layer, and garbage bags on the second. (Subtask #1) The left arm locates and grasps the drawer handle on the first floor with the, and the right arm is raised to the position where the garbage bag can be grasped. (Subtask #2) The left arm opens the drawer and keeps it open. The right arm picks up the trash bag and brings it to the top of the first drawer layer. (Subtask #3) After the right arm places the garbage bag in the middle of the drawer, it locates the medicine. (Subtask #4) Robot locates the top drawer on the left, then grabs the medicine with the robot’s right arm. (Subtask #5) It pulls the drawer open with the left arm, lifts the right arm to the top of the drawer, and places it in the middle. (Subtask #6) The left arm closes the drawer, and the right arm returns to the initial position.
(6)
“Put the specified fruit into the box”. There are boxes on the left side of the table, and bananas and apples on the right. (Subtask #1) The right arm is instructed to choose whether to grab an apple or a banana. (Subtask #2) The right arm places the specified fruit into the box.
(7)
“Store bowls, chopsticks, and spoons in shelves and drawers”. Bowls with chopsticks and spoons inside are placed on the table. (Subtask #1) Robot positions the left arm and grabs the handle of the first drawer. (Subtask #2) Robot opens the drawer with the left arm and keeps it open. It positions the right arm above the chopstick and grabs it. (Subtask #3) It puts the chopsticks on the top shelf with the right arm. (Subtask #4) Then it moves the right arm over the spoon in the bowl. The spoon is easy to move in the bowl, so it needs to be grasped slowly. (Subtask #5) After picking up the spoon with its right arm, the robot moves it to the top of the drawer. (Subtask #6) The right arm needs to accurately place the spoon into the drawer and withdraw the right arm. (Subtask #7) Robot closes the drawer with the left arm, moves the right arm to the position where the bowl can be grasped, and grasps the bowl. (Subtask #8) The left arm closes the drawer, withdraws, and returns to the initial position, and the right arm picks up the bowl and delivers it to the first position on the shelf. (Subtask #9) The right arm releases the mechanical claw to lower the bowl and then withdraws to the initial position.
Figure 5. Task 1: put the red square into the drawer [Supplementary Materials: Video S1].
Figure 5. Task 1: put the red square into the drawer [Supplementary Materials: Video S1].
Sensors 26 00165 g005
Figure 6. Task 2: pour water.
Figure 6. Task 2: pour water.
Sensors 26 00165 g006
Figure 7. Task 3: spoon into cup.
Figure 7. Task 3: spoon into cup.
Sensors 26 00165 g007
Figure 8. Task 4: surface cleaning.
Figure 8. Task 4: surface cleaning.
Sensors 26 00165 g008
Figure 9. Task 5: classified storage of materials.
Figure 9. Task 5: classified storage of materials.
Sensors 26 00165 g009
Figure 10. Task 6: sorting fruit.
Figure 10. Task 6: sorting fruit.
Sensors 26 00165 g010
Figure 11. Task 7: arrangement and placement of utensils.
Figure 11. Task 7: arrangement and placement of utensils.
Sensors 26 00165 g011aSensors 26 00165 g011b

4.2. Comparison of Experimental Results

In this paper, we compared the proposed method with three mainstream baseline methods—RDT [33], OpenVLA [34], and Aloha [35]. Our model was trained in a PyTorch Version 2.1.0 environment with an 8 GB NVIDIA RTX 4060 graphics card. We increased the model parameter size to 500 MB and the pre-trained model size to 122 MB.
Then, we applied the pre-trained model to our model for 10,000 iterations. Pretrained large language model batch size is 512, MemVLA model batch size is 1024, Optimizer: AdamW, pretrained large language model learning rate is 1 × 10−4, MemVLA model learning rate is 1 × 10−3, dimension size is 1024. The MemVLA model loss function is L1loss.
We adopted the success rate as the main metric, defined as the number of successful trials divided by the total number of trials. Each scenario was tested 25 times. To understand the model’s capability boundaries and error types during actual execution, we decomposed each task into multiple operational sub-stages and analyzed the model’s performance on these key sub-stages. This evaluation method not only focused on the overall task’s success but also emphasized fine-grained decision-making throughout the process. Specifically, each task typically consists of an object recognition phase, grasp phase, operation or transmission phase, fine interaction stage.
Object recognition phase: Locate the task goal, such as finding the correct drawer, item, or target container. Grasp phase: The manipulator is controlled to reach the target position and complete the firm grasp. Operation or transmission phase: The grasped object is moved to the target position, and the path rationality and attitude are controlled. Fine interaction stage: high-precision interactive actions such as inserting, dumping, and closing the drawer, shown in Table 1, Table 2, Table 3, Table 4, Table 5, Table 6 and Table 7.
To verify the advantages of the proposed lightweight self-attention mechanism in terms of resource efficiency, we systematically compared it with the standard fully connected self-attention method under a completely consistent training environment (including hardware platform, optimizer configuration, batch size, and training data). The evaluation metrics include GPU memory usage and the average time per training round.
The proposed method greatly speeds up training while significantly reducing video memory usage, effectively alleviating the computational and memory bottlenecks of the self-attention mechanism in long-sequence scenarios. As shown in Table 8, we compare the memory usage and training delay of the two attention mechanisms under the same settings, which verifies the superiority of the lightweight design.
The end-effector trajectory diagrams for the proposed method and other approaches are shown in the following Figure 12, Figure 13, Figure 14, Figure 15, Figure 16, Figure 17 and Figure 18, which compares trajectories from OpenVLA, Aloha, and RDT in sequence. In the figure, the orange endpoints represent the trajectory endpoints output by existing SOTA methods, while the green endpoints denote that output by the proposed method. The solid lines indicate the motion trajectories of the proposed method. It should be noted that, in motor control, we did not perform interpolation; instead, the motors were commanded directly to reach key waypoints, resulting in solid lines. The arrows indicate the direction of the end effector’s gripper movement.
From the above analysis, it can be observed that state-of-the-art (SOTA) methods such as OpenVLA, Aloha, and RDT exhibit significant jitter in robotic arm task execution, with this phenomenon predominantly concentrated between defined key action points. The core of this issue lies in their inherent inference mechanism: analogous to the reasoning logic of large language models, these methods require sequential inference and decision-making for each individual action. Consequently, the generated action sequence a 0 , , a i , , a r struggles to maintain motion continuity. In contrast, the ideal trajectory of a robotic arm should be a smooth straight line or curve; thus, the discretely inferred action points fluctuate around it, leading to considerable trajectory variance.
Delving into the fundamental causes, jitter primarily stems from two core factors: first, the teleoperation bias at the dataset level. If teleoperators fail to skillfully and precisely control the robotic arm via the joystick, the collected action sequence data will inherently contain significant jitter. Subsequent deep learning models, when fitting such flawed data, further amplify this intrinsic defect. Second, the inherent limitations of deep learning itself. Models trained using deep learning can only approximate the collected motion trajectories to within finite error, rather than achieving perfect alignment. This approximation error ultimately manifests as jitter in actual robotic motion.
The MemVLA model introduced targeted improvements to the inference mechanism: the model only needs to infer and generate key action points. During execution, the robotic arm merely requires accurate positioning at these key nodes, while the action sequences between consecutive key points are autonomously executed by the robotic system without additional model inference. This design inherently suppressed robotic arm jitter at the root of the reasoning logic, ensuring smooth trajectories.

5. Conclusions

This paper proposed a visual–language action-generation model for fine-grained manipulation tasks. By combining an efficient dual-arm collaboration strategy with a lightweight memory-gated filtering attention mechanism, a multimodal task dataset covering seven typical two-hand fine manipulation tasks was constructed. Through detailed evaluation at the task and sub-stage levels, we demonstrated the model’s stability and versatility in handling multi-stage collaborative tasks. We can apply the trained multi-head self-attention mechanism to the automatic handling of automotive interior parts in practice.
Our training method offered significant advantages in data efficiency, the autoregressive generation strategy, and task completion rate. Specifically, with memory-gated filtering attention, we have reduced the memory footprint by 72% and sped up training by an order of magnitude (from 1.35 s per batch to 0.129 s) while maintaining higher motion accuracy and robustness during critical task phases. The proposed VLA model could provide a more efficient and scalable technical path for large-scale, high-complexity robot task execution in future.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/s26010165/s1, Video S1: Put up the red square.

Author Contributions

All authors made substantial contributions to the manuscript. Conceptualization, B.R. and D.S.; Methodology, B.R. and D.S.; Data acquisition, D.S.; Data analysis and interpretation, B.R. and D.S.; Manuscript drafting, D.S.; Critical revision of the manuscript, B.R.; Funding acquisition, B.R. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the National Natural Science Foundation of China (Grant No. 51775325); Joint Funds of the National Natural Science Foundation of China (Grant No. U21A20121); Key Research and Development Program of Ningbo (Grant No. 2023Z218).

Institutional Review Board Statement

The study was conducted according to the guidelines of the Declaration of Helsinki and approved by the Academic Ethics Committee of Shanghai University (2022-011), data of approval: 22 February 2022.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

All data included in this study is available upon request by contact with the corresponding author. Some collected data is not publicly available because of ethical restrictions.

Acknowledgments

The authors appreciate for the support from the cooperating company, SAIC Motor R&D Innovation Headquarters, SAIC Motor Corporation Limited, Shanghai 201804, China. The authors also thank to the laboratory members Yongfeng Geng, Yunxiang Wang, Menghan Wu, Ruicheng Zhang and Quanxi Liao for their support in experimental data acquisition of humanoid robot.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Moldovan, B.; Moreno, P.; van Otterlo, M.; Santos-Victor, J.; De Raedt, L. Learning Relational Affordance Models for Robots in Multi-Object Manipulation Tasks. In Proceedings of the IEEE International Conference on Robotics and Automation (ICRA), Saint Paul, MN, USA, 14–18 May 2012; pp. 4373–4378. [Google Scholar]
  2. Duan, K.K.; Zou, Z.B.; Yang, T.Y. Training of construction robots using imitation learning and environmental rewards. Comput.-Aided Civ. Infrastruct. Eng. 2025, 40, 1150–1165. [Google Scholar] [CrossRef]
  3. Liang, X.; Sheng, L.; Cai, J. Contexts Matter: Robot-Aware 3D human motion prediction for Agentic AI-empowered Human-Robot collaboration. Adv. Eng. Inform. 2025, 68, 103591. [Google Scholar] [CrossRef]
  4. Su, Y.; Zhan, X.; Fang, H.; Li, Y.-L.; Lu, C.; Yang, L. Motion Before Action: Diffusing Object Motion as Manipulation Condition. IEEE Robot. Autom. Lett. 2025, 10, 7428–7435. [Google Scholar] [CrossRef]
  5. Duan, A.; Batzianoulis, I.; Camoriano, R.; Rosasco, L.; Pucci, D.; Billard, A. A structured prediction approach for robot imitation learning. Int. J. Robot. Res. 2024, 43, 113–133. [Google Scholar] [CrossRef]
  6. Jung, D.; Lee, H.; Yoon, S. Sample-efficient adversarial imitation learning. J. Mach. Learn. Res. 2024, 25, 1–32. [Google Scholar]
  7. Shafiullah, N.M.; Cui, Z.J.; Altanzaya, A.; Pinto, L. Behavior Transformers: Cloning k modes with one stone. Adv. Neural Inf. Process. Syst. 2022, 35, 22955–22968. [Google Scholar]
  8. Jonnavittula, A.; Parekh, S.; Losey, D.P. View: Visual imitation learning with waypoints. Auton. Robot. 2025, 49, 5. [Google Scholar] [CrossRef]
  9. Jang, E.; Irpan, A.; Khansari, M.; Kappler, D.; Ebert, F.; Lynch, C.; Levine, S.; Finn, C. Bc-z: Zero-shot task generalization with robotic imitation learning. In Proceedings of the Conference on Robot Learning, London, UK, 11 January 2022. [Google Scholar]
  10. Ning, Y.M.; Li, T.J.; Zhang, Y.L.; Li, Z.; Du, W.Q.; Zhang, Y. An Integrated Framework of Grasp Detection and Imitation Learning for Space Robotics Applications. Chin. J. Mech. Eng. 2025, 38, 139. [Google Scholar] [CrossRef]
  11. Ke, L.; Wang, J.; Bhattacharjee, T.; Boots, B.; Srinivasa, S. Grasping with Chopsticks: Combating Covariate Shift in Model-free Imitation Learning for Fine Manipulation. In Proceedings of the 2021 IEEE International Conference on Robotics and Automation (ICRA 2021), Xi’an, China, 30 May–5 June 2021. [Google Scholar]
  12. Morrison, D.; Corke, P.; Leitner, J. Closing the loop for robotic grasping: A real-time, generative grasp synthesis approach. arXiv 2018, arXiv:1804.05172. [Google Scholar] [CrossRef]
  13. Bonsignorio, F.; Cervellera, C.; Macciò, D.; Zereik, E. An imitation learning approach for the control of a low-cost low-accuracy robotic arm for unstructured environments. Int. J. Intell. Robot. Appl. 2023, 7, 13–30. [Google Scholar] [CrossRef]
  14. von Hartz, J.O.; Welschehold, T.; Valada, A.; Boedecker, J. The Art of Imitation: Learning Long-Horizon Manipulation Tasks from Few Demonstrations. IEEE Robot. Autom. Lett. 2024, 9, 11369–11376. [Google Scholar] [CrossRef]
  15. Kim, J.W.; Chen, J.T.; Hansen, P.; Shi, L.X.; Goldenberg, A.; Schmidgall, S.; Scheikl, P.M.; Deguet, A.; White, B.M.; Tsai, D.R.; et al. SRT-H: A hierarchical framework for autonomous surgery via language-conditioned imitation learning. Sci. Robot. 2025, 10, 5254. [Google Scholar]
  16. Mon-Williams, R.; Li, G.; Long, R.; Du, W.; Lucas, C.G. Embodied large language models enable robots to complete complex tasks in unpredictable environments. Nat. Mach. Intell. 2025, 7, 592–601. [Google Scholar] [CrossRef]
  17. Bärmann, L.; Kartmann, R.; Peller-Konrad, F.; Niehues, J.; Waibel, A.; Asfour, T. Incremental learning of humanoid robot behavior from natural interaction and large language models. Front. Robot. AI 2024, 11, 1455375. [Google Scholar] [CrossRef]
  18. Raptis, E.K.; Kapoutsis, A.C.; Kosmatopoulos, E.B. Agentic LLM-based robotic systems for real-world applications: A review on their agenticness and ethics. Front. Robot. AI 2025, 12, 1605405. [Google Scholar] [CrossRef]
  19. Wang, J.; Shi, E.; Hu, H.; Ma, C.; Liu, Y.; Wang, X.; Yao, Y.; Liu, X.; Ge, B.; Zhang, S. Large language models for robotics: Opportunities, challenges, and perspectives. J. Autom. Intell. 2025, 4, 52–64. [Google Scholar] [CrossRef]
  20. Fan, J.M.; Yin, Y.; Wang, T.; Dong, W.H.; Zheng, P.; Wang, L.H. Vision-language model-based human-robot collaboration for smart manufacturing: A state-of-the-art survey. Front. Eng. Manag. 2025, 12, 177–200. [Google Scholar] [CrossRef]
  21. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, L.; Polosukhin, I. Attention is All You Need. In Proceedings of the Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; Volume 30. [Google Scholar]
  22. Levine, S.; Finn, C.; Darrell, T.; Abbeel, P. End-to-End Training of Deep Visuomotor Policies. J. Mach. Learn. Res. 2016, 17, 39. [Google Scholar]
  23. Khan, S.; Naseer, M.; Hayat, M.; Zamir, S.W.; Khan, F.S.; Shah, M. Transformers in Vision: A Survey. ACM Comput. Surv. 2022, 54, 200. [Google Scholar] [CrossRef]
  24. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An Image is Worth 16 × 16 Words: Transformers for Image Recognition at Scale. arXiv 2020, arXiv:2010.11929. [Google Scholar]
  25. Du, Y.; Jian, J.; Zhu, Z.; Pan, D.; Liu, D.; Tian, X. A syntactic method for robot imitation learning of complex sequence task. Robot. Intell. Autom. 2023, 43, 132–143. [Google Scholar] [CrossRef]
  26. Yang, S.; Zhang, W.; Song, R.; Lu, W.Z.; Wang, H.S.; Li, Y.B. Explicit-to-Implicit Robot Imitation Learning by Exploring Visual Content Change. IEEE-ASME Trans. Mechatron. 2022, 27, 4920–4931. [Google Scholar] [CrossRef]
  27. Tanwani, A.K.; Yan, A.; Lee, J.; Calinon, S.; Goldberg, K. Sequential robot imitation learning from observations. Int. J. Robot. Res. 2021, 40, 1306–1325. [Google Scholar] [CrossRef]
  28. Gao, C.K.; Gao, H.C.; Guo, S.Q.; Zhang, T.; Chen, F. CRIL: Continual Robot Imitation Learning via Generative and Prediction Model. In Proceedings of the 2021 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Prague, Czech Republic, 27 September–1 October 2021. [Google Scholar]
  29. Lee, J.N.; Laskey, M.; Tanwani, A.K.; Aswani, A.; Goldberg, K. Dynamic regret convergence analysis and an adaptive regularization algorithm for on-policy robot imitation learning. Int. J. Robot. Res. 2021, 40, 1284–1305. [Google Scholar] [CrossRef]
  30. Cong, M.; Jian, J.; Zou, Q.; Liu, D. Robot imitation learning method based on structural grammar. J. Huazhong Univ. Sci. Technol. Nat. Sci. 2021, 49, 97–102. [Google Scholar]
  31. Kim, H.; Ohmura, Y.; Kuniyoshi, Y. Goal-Conditioned Dual-Action Imitation Learning for Dexterous Dual-Arm Robot Manipulation. IEEE Trans. Robot. 2024, 40, 2287–2305. [Google Scholar]
  32. Gubbala, G.N.S.; Nagashima, M.; Mori, H.; Seong, Y.A.; Sato, H.; Niiyama, R.; Suga, Y.; Ogata, T. Augmenting Compliance with Motion Generation Through Imitation Learning Using Drop-Stitch Reinforced Inflatable Robot Arm with Rigid Joints. IEEE Robot. Autom. Lett. 2024, 9, 8595–8602. [Google Scholar] [CrossRef]
  33. Liu, S.; Wu, L.; Li, B.; Tan, H.; Chen, H.; Wang, Z.; Xu, K.; Su, H.; Zhu, J. Rdt-1b: A Diffusion Foundation Model for Bimanual Manipulation. arXiv 2024, arXiv:2410.07864. [Google Scholar] [CrossRef]
  34. Kim, M.J.; Pertsch, K.; Karamcheti, S.; Xiao, T.; Balakrishna, A.; Nair, S.; Rafailov, R.; Foster, E.; Lam, G.; Sanketi, P.; et al. Openvla: An open-source vision-language-action model. arXiv 2024, arXiv:2406.09246. [Google Scholar]
  35. Zhao, T.Z.; Kumar, V.; Levine, S.; Finn, C. Learning Fine-Grained Bimanual Manipulation with Low-Cost Hardware. arXiv 2023, arXiv:2304.13705. [Google Scholar]
Figure 1. The operation of Vision-Language-Action (VLA).
Figure 1. The operation of Vision-Language-Action (VLA).
Sensors 26 00165 g001
Figure 2. Memory-gated filtering attention model.
Figure 2. Memory-gated filtering attention model.
Sensors 26 00165 g002
Figure 3. The strategy is collected for key steps after grasping.
Figure 3. The strategy is collected for key steps after grasping.
Sensors 26 00165 g003
Figure 4. Visual action key steps data collection method.
Figure 4. Visual action key steps data collection method.
Sensors 26 00165 g004
Figure 12. Comparison of trajectories between MemVLA, OpenVLA, Aloha, and RDT in Task 1.
Figure 12. Comparison of trajectories between MemVLA, OpenVLA, Aloha, and RDT in Task 1.
Sensors 26 00165 g012
Figure 13. Comparison of trajectories between MemVLA, OpenVLA, Aloha, and RDT in Task 2.
Figure 13. Comparison of trajectories between MemVLA, OpenVLA, Aloha, and RDT in Task 2.
Sensors 26 00165 g013
Figure 14. Comparison of trajectories between MemVLA, OpenVLA, Aloha, and RDT in Task 3.
Figure 14. Comparison of trajectories between MemVLA, OpenVLA, Aloha, and RDT in Task 3.
Sensors 26 00165 g014
Figure 15. Comparison of trajectories between MemVLA, OpenVLA, Aloha, and RDT in Task 4.
Figure 15. Comparison of trajectories between MemVLA, OpenVLA, Aloha, and RDT in Task 4.
Sensors 26 00165 g015
Figure 16. Comparison of trajectories between MemVLA, OpenVLA, Aloha, and RDT in Task 5.
Figure 16. Comparison of trajectories between MemVLA, OpenVLA, Aloha, and RDT in Task 5.
Sensors 26 00165 g016
Figure 17. Comparison of trajectories between MemVLA, OpenVLA, Aloha, and RDT in Task 6.
Figure 17. Comparison of trajectories between MemVLA, OpenVLA, Aloha, and RDT in Task 6.
Sensors 26 00165 g017
Figure 18. Comparison of trajectories between MemVLA, OpenVLA, Aloha, and RDT in Task 7.
Figure 18. Comparison of trajectories between MemVLA, OpenVLA, Aloha, and RDT in Task 7.
Sensors 26 00165 g018
Table 1. Comparison of put the red square into the drawer task of utensils completion rates.
Table 1. Comparison of put the red square into the drawer task of utensils completion rates.
Open Upper Drawer PositionGrab Parts BoxPrecisely PlaceDrawer CloseTotal
OpeVLA (%)1000500
Aloha (%)5080107010
RDT (%)6888778965
MemVLA (%)100909010090
Table 2. Comparison of pour water task of utensils completion rates.
Table 2. Comparison of pour water task of utensils completion rates.
Locating Liquid ContainersGrab the Container and Move ItFill Up LiquidPut the Bottle BackTotal
OpeVLA (%)500000
Aloha (%)88100377337
RDT (%)100100778977
MemVLA (%)1001009010090
Table 3. Comparison of spoon into cup task of utensils completion rates.
Table 3. Comparison of spoon into cup task of utensils completion rates.
Grab the SpoonPlace the Spoon into the ContainerStable WithdrawalTotal
OpeVLA (%)74407440
Aloha (%)1008010080
RDT (%)9010010090
MemVLA (%)1009510095
Table 4. Comparison of surface cleaning task of utensils completion rates.
Table 4. Comparison of surface cleaning task of utensils completion rates.
Identify and Locate the PaperPull Out a Paper TowelWipe the TableTotal
OpeVLA (%)80000
Aloha (%)100308030
RDT (%)100457045
MemVLA (%)1008010080
Table 5. Comparison of classified storage of materials task of utensils completion rates.
Table 5. Comparison of classified storage of materials task of utensils completion rates.
Grab Garbage BagOpen the First DrawerPlace to the First DrawerClose the First DrawerIdentify and Grab MedicineOpen the Second DrawerPlace to the Second DrawerClose the Second DrawerTotal
OpeVLA (%)702005050100500
Aloha (%)806015708750127012
RDT (%)8055771008867509055
MemVLA (%)100100851001001009210085
Table 6. Comparison of sorting fruit task of utensils completion rates.
Table 6. Comparison of sorting fruit task of utensils completion rates.
Grab the Fruit CorrectlyPick Up the Fruit CorrectlyPlace the Fruit CorrectlyTotal
OpeVLA (%)501008050
Aloha (%)////
RDT (%)901009090
MemVLA (%)100100100100
Table 7. Comparison of arrangement and placement task of utensils completion rates.
Table 7. Comparison of arrangement and placement task of utensils completion rates.
Open the First DrawerIdentify the Grasping ChopsticksPlace on the First ShelfIdentify the Grasping SpoonPut It in a DrawerClose the DrawerIdentify the Grasping BowlPlace on the Second ShelfTotal
OpeVLA (%)2560050204000
Aloha (%)638050201560907015
RDT (%)608070407590908060
MemVLA (%)100100887090100100100100
Table 8. Memory usage comparison.
Table 8. Memory usage comparison.
Self-AttentionMemory-Gated Filtering-Attention
Memory Usage (mb)13,6083754
Training time (s)1.350.129
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Ren, B.; Shi, D. Cross-Modality Alignment Perception and Multi-Head Self-Attention Mechanism for Vision-Language-Action of Humanoid Robot. Sensors 2026, 26, 165. https://doi.org/10.3390/s26010165

AMA Style

Ren B, Shi D. Cross-Modality Alignment Perception and Multi-Head Self-Attention Mechanism for Vision-Language-Action of Humanoid Robot. Sensors. 2026; 26(1):165. https://doi.org/10.3390/s26010165

Chicago/Turabian Style

Ren, Bin, and Diwei Shi. 2026. "Cross-Modality Alignment Perception and Multi-Head Self-Attention Mechanism for Vision-Language-Action of Humanoid Robot" Sensors 26, no. 1: 165. https://doi.org/10.3390/s26010165

APA Style

Ren, B., & Shi, D. (2026). Cross-Modality Alignment Perception and Multi-Head Self-Attention Mechanism for Vision-Language-Action of Humanoid Robot. Sensors, 26(1), 165. https://doi.org/10.3390/s26010165

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Article metric data becomes available approximately 24 hours after publication online.
Back to TopTop