Next Article in Journal
State-of-the-Art Review of Structural Vibration Control: Overview and Research Gaps
Previous Article in Journal
Electroluminescent Sensing Coating for On-Line Detection of Zero-Value Insulators in High-Voltage Systems
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

MPVT: An Efficient Multi-Modal Prompt Vision Tracker for Visual Target Tracking

1
School of Computer Science and Engineering, University of Electronic Science and Technology of China, Chengdu 611731, China
2
Chengdu Union Big Data Tech. Inc., Chengdu 610041, China
3
Zhuhai Yiyuan Technology Co., Ltd., Zhuhai 519040, China
4
Suining Institute of Digital Economy, Suining 629018, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2025, 15(14), 7967; https://doi.org/10.3390/app15147967
Submission received: 18 June 2025 / Revised: 11 July 2025 / Accepted: 14 July 2025 / Published: 17 July 2025
(This article belongs to the Special Issue Advanced Technologies Applied for Object Detection and Tracking)

Abstract

Visual target tracking is a fundamental task in computer vision. Combining multi-modal information with tracking leverages complementary information, which improves the precision and robustness of trackers. Traditional multi-modal tracking methods typically employ a full fine-tuning scheme, i.e., fine-tuning pre-trained single-modal models to multi-modal tasks. However, this approach suffers from low transfer learning efficiency, catastrophic forgetting, and high cross-task deployment costs. To address these issues, we propose an efficient model named multi-modal prompt vision tracker (MPVT) based on an efficient prompt-tuning paradigm. Three key components are involved in the model: a decoupled input enhancement module, a dynamic adaptive prompt fusion module, and a fully connected head network module. The decoupled input enhancement module enhances input representations via positional and type embedding. The dynamic adaptive prompt fusion module achieves efficient prompt tuning and multi-modal interaction using scaled convolution and low-rank cross-modal attention mechanisms. The fully connected head network module addresses the shortcomings of traditional convolutional head networks such as inductive biases. Experimental results from RGB-T, RGB-D, and RGB-E scenarios show that MPVT outperforms state-of-the-art methods. Moreover, MPVT can save 43.8% GPU memory usage and reduce training time by 62.9% compared with a full-parameter fine-tuning model.

1. Introduction

Visual target tracking is widely used in various scenarios, such as video surveillance and autonomous driving. Target tracking based on RGB images [1,2,3] has high performance due to its large-scale training data. However, such images often make tracking fail under poor lighting conditions, such as darkness, cluttered backgrounds, and high-speed motion. With the development of multi-modal technology, auxiliary modalities can enhance the success rate, precision, and robustness of tracking by utilizing the complementarity between modalities. Some classical scenarios include RGB-T tracking [4,5,6], RGB-D tracking [7,8], and RGB-E tracking [9,10,11].
For RGB-T tracking, some researchers have proposed fusion methods based on feature decoupling. Li et al. [12] constructed a multi-branch network to handle RGB and thermal infrared images separately. ADRNet [13] further utilizes adaptive universal branches for decoupling. MANet [14] jointly learns shared, modal-specific, and instance-aware features through universal, modal, and instance adapters. MANet++ [15] incorporates a hierarchical difference loss to maximize the distribution difference between modal-specific and shared features. DMCNet [16] adopts a dual-door control mutual condition module to suppress noise through cross-modal feature guidance. With the deepening development of attention mechanisms, some researchers have proposed dynamic fusion methods based on feature selection. For example, FANet [17] involves a hierarchical feature aggregation module to adaptively fuse multi-layer features. CMPP [18] includes a cross-modal propagation framework that achieves complementarity through spatio-temporal domain feature diffusion. ViPT [19] uses infrared modalities as visual prompts for visible modalities and utilizes the attention mechanism to achieve cross-modal interaction. In addition, some methods incorporate the unique characteristics of Transformer models to design multi-modal interaction modules for RGB-T target tracking. For example, MTNet [20] utilizes channel aggregation and distribution modules to eliminate redundancy. SiamCAF [21] uses complementary coupling modules to enhance feature similarity through cross-connections. Very recently, Wang et al. [22] proposed a novel lightweight adapter to boost fusion performance for RGB-T target tracking.
Traditional RGB-D target tracking methods mainly adopt manually designed features and shallow fusion strategies. For example, DS-KCF [23] achieves bimodal feature-level fusion through HOG feature extraction, while STC [24] improves tracking robustness through decision-level fusion. With the development of deep learning, researchers have begun to explore deep feature fusion. For example, Jiang et al. [25] proposed a dual-branch CNN architecture, which extracts RGB and deep features separately and enhances discriminability through weighted fusion. OTR [26] introduces deep dynamic spatial constraints to optimize DCF (discriminative correlation filter) learning, significantly improving tracking stability in occluded scenarios. In recent years, researchers have focused more on cross-modal interaction modeling to fully utilize the complementary information of multiple modalities. DeT [7] incorporates a symmetrical double-branch structure and achieves early fusion through feature concatenation. Recently, Gao et al. [27] proposed a new RGB-D approach to capture the interactive information between modalities.
The RGB-E tracking method [28] often uses a cross-modal attention mechanism for feature fusion. AFNet [10] utilizes the high temporal resolution of event data to design a cross-frame rate alignment module, achieving a tracking frequency of 1000 Hz. HRCEUTrack [11] introduces a Mask Autoencoder (MAE) to randomly mask RGB and event tokens, enhances cross-modal representation consistency through reconstruction tasks, and combines orthogonal high-rank regularization to suppress network fluctuations.
In recent years, the visual prompt learning paradigm has been introduced in target tracking, which adapts specific downstream tasks through prompt modules to fully utilize pre-trained knowledge and reduce dependence on downstream data. Yang et al. [29] proposed ProTrack, which applies prompt learning methods to multi-modal tracking. This model utilizes visual prompts to convert RGB-T and RGB-D data into a unified RGB space. With the diversification of tracking tasks, some researchers are focusing on a universal visual prompt framework. Hong et al. [30] proposed a OneTracker framework, which freezes the basic tracking model and only fine-tunes the cross-modal prompt module, achieving the unified processing of multiple tracking tasks, enabling the model to quickly adapt to diverse tracking scenarios. Due to target tracking being a temporal task, some methods rely on prompt learning to fuse spatio-temporal information. Shi et al. [31] proposed an EVPTrack method, which effectively conveys spatio-temporal information to the current frame by generating an explicit visual prompt, avoiding frequent template update operations. This method enhances the model’s adaptability to scale changes through a multi-scale prompt mechanism.
Although various methods have achieved great success in target tracking, they mainly adopt full-parameter fine-tuning strategies, which have some limitations: (1) Low transfer efficiency: The introduction of multi-modality further increases the complexity of a model, resulting in a significant increase in the resources required for fine-tuning. (2) Catastrophic forgetting of pre-trained knowledge: Unbalanced data sizes can easily lead to the over-fitting of downstream noise in a model. (3) High cost of cross-task deployment: Each downstream task model needs to be stored separately and cannot share common capabilities.
This paper focuses on the issues of multi-modal tracking, combining thermal infrared (T), depth (D), event (E), or other data from different sensors with RGB images, and makes full use of the complementary characteristics between modalities to enhance tracking performance. In this paper, a multi-modal efficient fine-tuning model based on prompt learning, MPVT, is proposed to address the problems of low transfer efficiency, catastrophic forgetting, and weak multi-modal correlation modeling in the traditional full-parameter fine-tuning transfer paradigm.
The remainder of this paper is organized as follows: In Section 2, the multi-modal prompt vision tracker (MPVT) is proposed based on an efficient prompt-tuning paradigm. Decoupled input enhancement, dynamic adaptive prompt fusion, and fully connected head network modules are included in the model. Furthermore, a joint loss function is introduced so as to train the model efficiently. The experimental settings, datasets, details of the experimental results, and analysis are presented in Section 3. Finally, conclusions are summarized in Section 4.

2. Materials and Methods

The framework of MPVT is shown in Figure 1. MPVT aims to fine-tune the pre-trained RGB target tracking model for multi-modal tracking tasks, including RGB-T, RGB-D, and RGB-E, through efficient transfer learning.
Based on the ViT encoder architecture, three components are designed to achieve efficient multi-modal transfer learning. (1) The decoupled input embedding enhancement module first enhances the multi-modal resolution-decoupled position embedding of inputs and then enhances type embedding for both modalities. (2) The dynamic adaptive prompt fusion module is the core of prompt learning, which uses N multi-modal prompt learning modules for transfer learning, while significantly reducing the number of learning parameters by freezing the backbone encoder layer. (3) The fully connected head network module implements two branches of classification and regression, avoiding the inductive bias of traditional convolutional head networks and demonstrating stronger global feature interaction capabilities.

2.1. Decoupled Input Embedding Enhancement Module

For the input template images Z R G B , Z A U X R H z × W z × 3 and search regions X R G B , X A U X R H x × W x × 3 , two image embedding layers are used to map them to token sequences. Since the backbone has been pre-trained, only the auxiliary modal embedding layer needs to be fine-tuned, while the RGB embedding layer is frozen. The embedding process can be written as
X ˜ R G B 0 = E R G B ( X R G B ) Z ˜ R G B 0 = E R G B ( Z R G B ) X ˜ A U X 0 = E A U X ( X A U X ) Z ˜ A U X 0 = E A U X ( Z A U X ) ,
where X ˜ R G B 0 , X ˜ A U X 0 R L x × d , Z ˜ R G B 0 , Z ˜ A U X 0 R L z × d , L x = H x s × W x s , and L z = H z s × W z s , with the patch size being ( s , s ) .
After obtaining the embedding token sequences for each image, this module enhances the input token using two procedures: resolution-decoupled positional embedding enhancement and type embedding enhancement.
Traditional multi-modal tracking models independently embed the positions of the template and searching regions by utilizing one-dimensional absolute embedding. If the resolution of the template region is significantly lower than that of the searching region, this will disrupt the input distribution of the pre-trained ViT model, resulting in the failure of fine-tuning. To solve this problem, unified position embedding through a resolution-decoupled position embedding method is presented. Specifically, let the one-dimensional absolute embedding be R ; the corresponding two-dimensional form is R 2 D , as shown in Equation (2). For a searching region of the size h x × w x , if h x = h a n d w x = w , R 2 D can be directly used as the position embedding of the search region, i.e., R x = R 2 D .
R 2 D = r 11 r 1 w r h 1 r h w ,
A sub-matrix of R x is used as the position embedding of a template region of the size h z × w z due to its resolution being smaller than that of the searching region. Specifically, starting from the top-left corner of R x , we select a region whose size equals the size of the template region, which can be written as
R z = r 11 r 1 w z r h z 1 r h z w z .
Since the position encoding of the template region R z and that of the searching region R x are derived from the same position encoding R 2 D , which is consistent with the pre-trained ViT model, the uniformity of position encoding can be ensured.
Additionally, two learnable embeddings, E z t y p e R d and E x t y p e R d , are designed to distinguish between the template and searching region embedding and between the targets in the template region and background embedding. E z t y p e is further divided into the target region embedding E z o t y p e R d and the background region embedding E z b t y p e R d . Through resolution-decoupled position and type embedding enhancement, the template region embedding E z and the searching region embedding E x can be represented as
E x ( i , j ) = R x ( i , j ) + E x t y p e ,
and
E z ( i , j ) = R z ( i , j ) + E z o t y p e , z ( i , j ) is the target region R z ( i , j ) + E z b t y p e , o t h e r w i s e ,
For this reason, the enhanced features can be obtained as
X R G B 0 = X ˜ R G B 0 + E x Z R G B 0 = z ˜ R G B 0 + E z X A U X 0 = X ˜ A U X 0 + E x Z A U X 0 = z ˜ A U X 0 + E z .

2.2. Dynamic Adaptive Prompt Fusion Module

As shown in Figure 1, we assume that there are N encoder layers, M 1 , , M N , in the ViT backbone and the corresponding N multi-modal prompt learning modules P 1 , , P N . The prompt learning process is
C 0 = C o n c a t ( X R G B 0 , Z R G B 0 ) C A U X 0 = C o n c a t ( X A U X 0 , Z A U X 0 ) C A U X i = P i ( C i 1 , C A U X i 1 ) C i = M i ( C i 1 + C A U X i ) ,
where Concat represents the concatenation operation of the tokens C i , C A U X i R L × d , and L = L x + L z . P i is the i-th dynamic adaptive prompt fusion module, which mainly includes three core components: (1) dynamic channel scaling convolution for modality-specific weighting, (2) low-rank cross-modal attention for multi-modal interaction, and (3) grouped linear mapping for efficient output mapping, as shown in Figure 2.
(1)
Dynamic Channel Scaling Convolution
For the modal feature C , the statistical value s of the feature is calculated by
s = 1 N i = 1 N C [ i , : ] R d ,
and the dynamic scaling factor α is calculated by
α = S o f t m a x ( W 2 · G E L U ( W 1 s ) ) ,
where W 1 R r × d and W 2 R K × r are learnable parameters, and K is the number of dynamic convolution kernels.
Two types of features are separately learned using depth-wise separable convolution through
X = R e s h a p e ( C ) R d × H × W M = D W C o n v ( X ) k = 1 K α [ k ] · C k ,
where D W C o n v : R d × H × W R ( d / β ) × H × W is a depth-wise separable convolution, C k R ( d / β ) × 1 × 1 is a dynamic channel scaling vector, and β is a hyper-parameter of the dimensionality reduction ratio.
The intermediate results M and M A U X can be obtained by implementing the above dynamic channel scaling convolution on two features, C i 1 and C A U X i 1 .
(2)
Low-Rank Cross-Modal Attention
A low-rank attention mechanism for cross-modal learning is proposed to overcome the large number of parameters and the high complexity of traditional attention mechanisms. For M and M A U X , low rank projection can be obtained by
Q = M W Q R N × d K = [ M , M A U X ] W K R 2 N × d V = [ M , M A U X ] W V R 2 N × d ,
where W Q , W K , W V R ( d / β ) × d are learnable parameters. Based on Equation (11), a multi-modal interaction feature, O , can be achieved by fusing the attention mechanism and residual features dynamically, as follows:
A t t n = S o f t m a x Q K d R N × 2 N O = γ · ( A t t n V ) + ( 1 γ ) · M A U X ,
where the fusion coefficient γ is a learnable parameter.
(3)
Grouped Linear Mapping
Efficient grouped linear mapping is used to calculate prompt output features that in Equation (7), C A U X i R L × d , as follows:
C A U X i = GroupLinear ( O ) = Concat j = 1 g ( O [ : , j · d / g : ( j + 1 ) · d / g ] W j ) ,
where W j R ( d / g ) × ( d / g ) is the group learnable parameter and g is the number of groups. For the sake of computational efficiency, g is taken as 8 in this study.

2.3. Fully Connected Head Network Module

Traditional models such as OSTrack [32] use convolutional head networks for classification and regression, which have strong inductive biases, such as locality assumptions and translation invariance. The local receptive field of convolutional kernels limits the global feature interaction ability. For this reason, a fully connected head network is utilized in this study, which entirely uses multi-layer perceptrons for classification and regression.
In this module, a multi-task joint loss function, L t o t a l = L c l s + L b o x , is used to train the model. The classification loss L c l s adopts binary cross-entropy loss (BCE) and can be defined by
L c l s = 1 N p o s i , j [ y ¯ i , j log y ^ i , j ( 1 y ¯ i , j ) log ( 1 y ^ i , j ) ] ,
where y ¯ i , j is the IoU value of the predicted box B i , j p r e d and the true label B g t for each foreground point, ( i , j ) . y ^ i , j = C [ i , j ] [ 0 , 1 ] is the predicted score of the classification branch, and N p o s is the total number of foreground points.
The regression loss L b o x adopts generalized IoU (GIoU), which can provide effective gradients even if the predicted and label boxes do not overlap. It can be defined as
L b o x = 1 N p o s i , j II y i , j = 1 1 GIoU ( B i , j p r e d , B g t ) ,
where II is the indicator function, which only calculates the loss of foreground points.

3. Experimental Results and Discussion

3.1. Datasets

The performance of MPVT was evaluated and compared using three multi-modal datasets, LasHeR [33], DepthTrack [7], and VisEvent [9], corresponding to three multi-modal scenarios: RGB-T, RGB-D, and RGB-E, respectively. LasHeR is the largest and most diverse publicly available dataset in RGB-T target tracking, consisting of 1224 visible light and thermal infrared video pairs with over 730K frames. DepthTrack includes 200 video sequences, with an average sequence length of 1473 frames, providing sufficiently long and continuously changing training data. In addition, it covers 40 different scenario types and 90 target categories. VisEvent contains 820 video pairs, which were collected in complex real-world scenarios such as those with low light, high-speed motion, and cluttered backgrounds. It is widely used for the performance evaluation of RGB-E tracking methods.

3.2. Evaluation Indices

Four evaluation indices are used to evaluate the performance of the methods, that is, the success rate ( S r ), the tracking recall rate ( T R e ), precision ( P r ), and the F1-score.
The success rate is the ratio of successfully tracked frames to the total number of frames at a specific IoU threshold, β . If the IoU value of a frame is greater than or equal to β , the tracking of that frame is successful. S r can be defined as
S r = 1 N i = 1 N I ( I o U i β ) ,
I o U i = | p i g i | | p i g i | ,
where N is the total number of frames in a video, I o U i is the IoU value of the i-th frame, and I ( · ) is the indicator function whose value is 1 if the condition is satisfied and 0 otherwise. p i is the predicted target position, box and g i is the label position box.
Precision measures the degree of overlap between predicted and real bounding boxes. It can be defined as
P r = 1 N i = 1 N I o U i ,
The tracking recall rate ( T R e ) is also used as an evaluation index to evaluate the tracking method. Unlike traditional recall, T R e measures the tracker’s ability to correctly detect targets in visible frames. It can be defined as
T R e = t = 1 N I ( v ^ t > β , v t = 1 ) t = 1 N I ( v t = 1 ) ,
where v t { 0 , 1 } is an annotation variable that determines whether the target is visible in frame t, and v ^ t [ 0 , 1 ] is the confidence score for frame t.
Generally, a large P r will lead to a low T R e and vice versa. For a better comparison, a unified index, i.e., the F1-score, combining P r and T R e is introduced to evaluate the performance of a method. It can be defined as
F 1 = 2 × P r · T R e P r + T R e .

3.3. Model Parameter Settings

The model was trained for 60 epochs with a batch size of 16; each batch contained 6 × 10 5 pairs of data. AdamW was used as the optimization algorithm, with the weight decay being 10 4 . The learning rate was updated by the CosineLR scheduling algorithm [34], with the maximal and minimal learning rates being 10 4 and 10 6 , respectively. The parameters were set as frozen or learnable for different modules, as shown in Figure 1. The learnable prompt parameters were initialized using the Xavier algorithm [35], and other training parameters are shown in Table 1.

3.4. Main Results

Various methods were compared with MPVT using three datasets, including TBSI [4], APFNet [5], FANet [17], CAT [12], MDNet [36], PrDIMP50 [37], LTMU [38], OSTrack [32], TransT [39], DAL [8], TATrack [40], TABBTrack [41], ProTrack [29], ViPT [19], and Un-Track [42].
For the RGB-T scenario, various types of RGB-T tracking models were compared with MPVT. Early methods such as FANet [5] mainly improved the complementarity of multi-modal information through feature fusion. Feature decoupling-based methods such as CAT [12] and APFNet [5] introduced difficult attribute branches and progressive fusion mechanisms, respectively. TransT [39] applied Transformer to visual tracking, significantly enhancing the global information interaction between templates and searching regions. OSTrack [32] and TBSI [4] further optimized the Transformer architecture, achieving significant improvements in precision and generalization performance. ProTrack [29] utilized a multi-modal prompt tracking framework, and ViPT [19] further fine-tuned the pre-trained model through visual prompts. We conducted extensive comparisons based on the various methods mentioned above, and the experimental results are shown in the third and fourth columns of Table 2. MPVT achieved the best performance among all models, with a success rate of 0.578 and a precision of 0.724. Compared with the second best model, TATrack [40], it increased the success rate and precision by 1.9% and 2.2%, respectively, and compared with ViPT [19], it increased the success rate and precision by 5.3% and 7.3%, respectively.
For the RGB-D scenario, to conduct a comprehensive comparison, various models were introduced as baselines. Besides the methods used in the RGB-T scenario, some other methods were also compared with MPVT. DAL [8] constructed a long-term tracking framework for deep perception through an end-to-end deep feature fusion network. TABBTrack [41] adopted a three-stream architecture and combined temporal features to fuse depth images. The results are shown in columns 5–7 in Table 2. MPVT achieved the highest performance among all the compared models, with an F1-score of 0.628, a tracking recall rate of 0.626, and a precision of 0.631. Compared with ViPT [19], it increased the F1-score, T R e , and P r by 3.4%, 3%, and 3.0%, respectively, and compared with the second-best model, TABBTrack [41], it increased the F1-score, T R e , and P r by 1%, 1.1%, and 0.9%, respectively.
Among all the compared models in the RGB-E scenario, MPVT achieved the highest tracking success rate (0.624) and precision (0.761), with improvements of over 3.2% and 0.3%, respectively, compared with those of the second-best model, ViPT [19], in the RGB-E scenario, as shown in the eighth and ninth columns of Table 2.
Regarding degraded scenarios, Figure 3 and Figure 4 present precision plots for 15 degraded scenarios involving the LasHeR and VisEvent datasets, respectively (DepthTrack is excluded due to unavailable degradation labels). MPVT consistently outperformed baselines across all scenarios. The performance gain was notable under abrupt illumination variation (AIV), hyaline occlusion (HO), and background clutter (BC) in LasHeR and also under low illumination (LI), AIV, and HO in the VisEvent dataset. For instance, MPVT outperformed ViPT by 0.306 under AIV. These results demonstrate the model’s robust performance in degraded conditions.
Regarding efficiency, MPVT has 0.9% (2.8 M) trainable parameters while achieving competitive tracking performance. It shows a higher transfer efficiency compared with traditional fully fine-tuning models, whose trainable parameter ratios are 100%, and learnable parameters are generally greater than 100 M. ViPT is a prompt-based model and has 0.4% (0.84 M) trainable parameters, and although MPVT has more trainable parameters than this, it achieved better tracking performance across all three datasets. Experimental results across the RGB-T, RGB-D, and RGB-E scenarios demonstrate that MPVT preserves the feature extraction capabilities of pre-trained models. Through decoupled input enhancement and dynamic prompt fusion, it avoids feature degradation while effectively integrating complementary multi-modal information. Compared with full-parameter fine-tuning, MPVT maintains higher transfer efficiency.

3.5. Ablation Study

To analyze the effectiveness of the proposed components, ablation experiments were conducted on the three datasets. The experimental results are shown in Table 3. MPVT, due to the use of efficient training paradigms such as prompt learning, achieved the best performance in RGB-T, RGB-D, and RGB-E, although only 0.9% of the parameters were used for training. First, if the fully connected head network module was removed and the traditional convolutional head network was used instead (MPVT-FC), the performance decreased by 0.01 to 0.02. This indicates that fully connected head networks can effectively avoid the inductive bias of traditional convolutional head networks. Second, if the dynamic adaptive prompt fusion module was removed based on MPVT-FC (MPVT-PF), the model just utilized a full-parameter fine-tuning strategy. In this case, each index decreased by 0.04 to 0.05 on all three datasets. This indicates that the fusion module is crucial for efficient parameter learning and can efficiently transfer from a single-modal pre-trained model on large-scale RGB data to multi-modal scenarios such as RGB-T, RGB-D, and RGB-E. Finally, if the input enhancement module was further removed (MPVT-IE), the performance decreased by about 0.01, proving that the proposed position and type embedding enhancement can contribute to accurate target tracking.

3.6. Model Analysis

3.6.1. Sensitivity Analysis of Prompt Depth

The dynamic adaptive prompt fusion (DAPF) module can be inserted before any ViT encoder layer. The number of DAPF modules used determines the prompt depth: shallow integration applies DAPF only in the first layer, whereas deep integration applies it in every layer. To further assess the impact of prompt fusion depth on tracking performance, we conducted experiments using a scaled-down MPVT model (with 12 encoder layers and 768 hidden dimensions). Performance variations were observed through primary metrics across the three datasets.
This experiment evaluated tracking performance at prompt depths N p { 1 , 3 , 5 , 9 , 12 } . As shown in Figure 5, increasing the prompt depth consistently enhanced performance across all three datasets, demonstrating the effectiveness and scalability of our dynamic adaptive prompt fusion (DAPF) module.

3.6.2. Feature Visualization

A visualization of the tracking results of MPVT and a baseline model in three multi-modal scenarios is shown in Figure 6. From Figure 6d,e, it can be seen that the feature map of MPVT is more concentrated, clearer, and less noisy, while that of the baseline model is more scattered. This indicates that MPVT has a more stable discrimination of targets. From the tracking results in Figure 6b,c, it can also be observed that MPVT (red box) has higher tracking precision compared with the baseline model (green box).

3.6.3. Model Efficiency

The efficiency of MPVT was analyzed from three aspects: training time, GPU memory usage, and inference speed, as shown in Table 4. The training time and GPU memory usage by prompt learning decreased significantly compared with those of full-parameter fine-tuning models, saving 43.8% of GPU memory and 62.9% of training time. In terms of inference speed, due to the addition of model parameters and computational complexity, the inference speed slightly decreased by about 1 fps. The dynamic adaptive prompted fusion (DAPF) module maintained consistent computational complexity for all auxiliary modalities. This minimal inference performance difference enables significant advantages in practice. Regarding edge deployment, 43.8% memory reduction allows operation on resource-constrained devices, and efficiency gains enable the simultaneous processing of more streams.

3.6.4. Differences Between MPVT and Vision–Language Prompting

MPVT shares core principles with CLIP-style prompting [43] but differs in three aspects: (1) Token construction: CLIP uses discrete text tokens, while MPVT employs continuous visual prompts requiring no vocabulary constraints. (2) Fusion granularity: CLIP fuses global semantic features, whereas our low-rank cross-attention preserves spatial details critical for localization. (3) Knowledge transfer: Both leverage frozen backbones, but MPVT’s dynamic scaling adapts to sensor-specific noise patterns unseen in CLIP. These differences highlight how prompt engineering must be rethought for dense prediction tasks.

4. Conclusions

Focusing on the issues of low transfer efficiency, catastrophic forgetting of pre-trained knowledge, and weak multi-modal correlation modeling in traditional full-parameter fine-tuning, an efficient multi-modal prompt vision tracker is proposed to achieve efficient prompt learning on the basis of an RGB pre-trained model. The proposed model can minimize the cost of fine-tuning downstream tasks while achieving higher multi-modal tracking performance. The model mainly includes three key components: a decoupled input enhancement module, a dynamic adaptive prompt fusion module, and a fully connected head network module. The decoupled input enhancement module enhances input from position and type embedding. The dynamic adaptive prompt fusion module achieves efficient prompt learning and multi-modal interaction through scaling convolution and low-rank cross-modal attention learning. The fully connected head network module overcomes the shortcomings of traditional convolutional head networks, such as inductive biases.
The performance of MPVT was evaluated under three multi-modal scenarios: RGB-T, RGB-D, and RGB-E. The experimental results show that the model achieved the best performance on all three multi-modal datasets, achieving tracking precisions of 72.4%, 63.1%, and 76.1%, respectively, and outperforming the second-best model by 2.2%, 0.9%, and 0.3%, respectively. Further ablation studies showed that all three proposed components contribute to more accurate target tracking. Furthermore, the model achieved efficient multi-modal transfer with less than 1% of learnable parameters. The efficiency analysis showed that MPVT is much more effective in training time and GPU memory usage than full-parameter fine-tuning models, saving up to 43.8% of GPU memory and 62.9% of training time.
Recently, many researchers have fused vision and text in various large models [44]. Extending MPVT to vision–language tracking requires solving the fundamental challenge of spatial–semantic alignment, i.e., bridging textual descriptions with visual coordinates while maintaining real-time efficiency. Key research directions include the following: (1) developing geometric-aware attention mechanisms to ground linguistic concepts (e.g., “left”, “occluded”) to pixel-level locations; (2) designing memory-augmented architectures that preserve object identity across evolving text descriptions; and (3) designing efficient fusion schemes to avoid computational explosion when combining visual tokens with linguistic tokens. In future work, we will introduce lingual information into MPVT to further extend multi-modal scenarios.

Author Contributions

Conceptualization, J.X. and D.C.; methodology, J.X., Y.F. (Yan Fu) and J.Z.; software, J.X. and T.H.; validation, T.H., X.W. and Y.F. (Yuke Fang); formal analysis, J.X., T.H. and X.W.; investigation, J.X., Y.F. (Yuke Fang) and D.C.; resources, X.W. and Y.F. (Yuke Fang); data curation, X.W. and Y.F. (Yuke Fang); writing—original draft preparation, J.X. and D.C.; writing—review and editing, J.X., J.Z., Y.F. (Yan Fu), X.W. and D.C.; visualization, T.H. and X.W.; supervision, J.Z. and D.C.; funding acquisition, Y.F. (Yan Fu) and D.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research was partially funded by the Major Science and Technology Projects in Sichuan Province under Grant No. 2024ZDZX0021 and the Major Program of National Natural Science Foundation of China under Grant No. T2293771.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Three datasets used in this paper are available from public URLs. LasHeR is available at https://github.com/BUGPLEASEOUT/LasHeR (accessed on 15 January 2025), DepthTrack is available at https://github.com/xiaozai/DeT (accessed on 21 January 2025), and VisEvent is available at https://github.com/wangxiao5791509/VisEvent_SOT_Benchmark (accessed on 2 February 2025).

Conflicts of Interest

Authors Yan Fu, Junlin Zhou, Tianxiang He, Yuke Fang and Duanbing Chen were employed by the company Chengdu Union Big Data Tech. Inc. Author Xiaopeng Wang was employed by the company Zhuhai Yiyuan Technology Co., Ltd. Author Duanbing Chen was employed by the agency Suining Institute of Digital Economy. The remaining authors declare that this research was conducted in the absence of any commercial or financial relationships that could be construed as potential conflicts of interest.

References

  1. Cui, Y.; Jiang, C.; Wang, L.; Wu, G. MixFormer: End-to-End Tracking with Iterative Mixed Attention. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 13598–13608. [Google Scholar] [CrossRef]
  2. Zhang, Z.; Peng, H. Deeper and Wider Siamese Networks for Real-Time Visual Tracking. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 4586–4595. [Google Scholar] [CrossRef]
  3. Chen, X.; Peng, H.; Wang, D.; Lu, H.; Hu, H. SeqTrack: Sequence to sequence learning for visual object tracking. In Proceedings of the 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Vancouver, BC, Canada, 17–24 June 2023; pp. 14572–14581. [Google Scholar] [CrossRef]
  4. Hui, T.; Xun, Z.; Peng, F.; Huang, J.; Wei, X.; Wei, X.; Dai, J.; Han, J.; Liu, S. Bridging search region interaction with template for RGB-T tracking. In Proceedings of the 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Vancouver, BC, Canada, 17–24 June 2023; pp. 13630–13639. [Google Scholar] [CrossRef]
  5. Xiao, Y.; Yang, M.; Li, C.; Liu, L.; Tang, J. Attribute-based progressive fusion network for RGBT tracking. Proc. AAAI Conf. Artif. Intell. 2022, 36, 2831–2838. [Google Scholar] [CrossRef]
  6. Feng, M.; Su, J. Learning reliable modal weight with transformer for robust RGBT tracking. Knowl.-Based Syst. 2022, 249, 108945. [Google Scholar] [CrossRef]
  7. Yan, S.; Yang, J.; Käpylä, J.; Zheng, F.; Leonardis, A.; Kämäräinen, J.K. DepthTrack: Unveiling the power of RGBD tracking. In Proceedings of the 2021 IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 10–17 October 2021; pp. 10705–10713. [Google Scholar] [CrossRef]
  8. Qian, Y.; Yan, S.; Lukežič, A.; Kristan, M.; Kämäräinen, J.K.; Matas, J. DAL: A deep depth-aware long-term tracker. In Proceedings of the 2020 25th International Conference on Pattern Recognition (ICPR), Milan, Italy, 10–15 January 2021; pp. 7825–7832. [Google Scholar] [CrossRef]
  9. Wang, X.; Li, J.; Zhu, L.; Zhang, Z.; Chen, Z.; Li, X.; Wang, Y.; Tian, Y.; Wu, F. VisEvent: Reliable object tracking via collaboration of frame and event flows. IEEE Trans. Cybern. 2024, 54, 1997–2010. [Google Scholar] [CrossRef] [PubMed]
  10. Zhang, J.; Wang, Y.; Liu, W.; Li, M.; Bai, J.; Yin, B.; Yang, X. Frame-event alignment and fusion network for high frame rate tracking. In Proceedings of the 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Vancouver, BC, Canada, 17–24 June 2023; pp. 9781–9790. [Google Scholar] [CrossRef]
  11. Zhu, Z.; Hou, J.; Wu, D.O. Cross-modal orthogonal high-rank augmentation for RGB-event transformer-trackers. In Proceedings of the 2023 IEEE/CVF International Conference on Computer Vision (ICCV), Paris, France, 1–6 October 2023; pp. 22045–22055. [Google Scholar]
  12. Li, C.; Liu, L.; Lu, A.; Ji, Q.; Tang, J. Challenge-aware RGBT tracking. In Computer Vision—ECCV 2020; Lecture Notes in Computer Science; Vedaldi, A., Bischof, H., Brox, T., Frahm, J.M., Eds.; Springer International Publishing: Cham, Switzerland, 2020; Volume 12367, pp. 222–237. [Google Scholar] [CrossRef]
  13. Zhang, P.; Wang, D.; Lu, H.; Yang, X. Learning adaptive attribute-driven representation for real-time RGB-T tracking. Int. J. Comput. Vis. 2021, 129, 2714–2729. [Google Scholar] [CrossRef]
  14. Li, C.L.; Lu, A.; Zheng, A.H.; Tu, Z.; Tang, J. Multi-adapter RGBT tracking. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision Workshop (ICCVW), Seoul, Republic of Korea, 27–28 October 2019; pp. 2262–2270. [Google Scholar] [CrossRef]
  15. Lu, A.; Li, C.; Yan, Y.; Tang, J.; Luo, B. RGBT tracking via multi-adapter network with hierarchical divergence loss. IEEE Trans. Image Process. 2021, 30, 5613–5625. [Google Scholar] [CrossRef] [PubMed]
  16. Lu, A.; Qian, C.; Li, C.; Tang, J.; Wang, L. Duality-gated mutual condition network for RGBT tracking. IEEE Trans. Neural Netw. Learn. Syst. 2025, 36, 4118–4131. [Google Scholar] [CrossRef] [PubMed]
  17. Zhu, Y.; Li, C.; Tang, J.; Luo, B. Quality-aware feature aggregation network for robust RGBT tracking. IEEE Trans. Intell. Veh. 2021, 6, 121–130. [Google Scholar] [CrossRef]
  18. Wang, C.; Xu, C.; Cui, Z.; Zhou, L.; Zhang, T.; Zhang, X.; Yang, J. Cross-modal pattern-propagation for RGB-T tracking. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 7062–7071. [Google Scholar] [CrossRef]
  19. Zhu, Y.; Guo, R.; Wang, Y.; Lu, W. ViPT: Visual prompt multi-modal tracking. In Proceedings of the 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Vancouver, BC, Canada, 17–24 June 2023; pp. 12234–12243. [Google Scholar]
  20. Hou, R.; Xu, B.; Ren, T.; Wu, G. MTNet: Learning modality-aware representation with transformer for RGBT tracking. In Proceedings of the 2023 IEEE International Conference on Multimedia and Expo (ICME), Brisbane, Australia, 10–14 July 2023; pp. 1163–1168. [Google Scholar] [CrossRef]
  21. Xue, Y.; Zhang, J.; Lin, Z.; Li, C.; Huo, B.; Zhang, Y. SiamCAF: Complementary attention fusion-based siamese network for RGBT tracking. Remote Sens. 2023, 15, 3252. [Google Scholar] [CrossRef]
  22. Wang, H.; Xu, T.; Tang, Z.; Wu, X.J.; Kittler, J. Multi-modal adapter for RGB-T tracking. Inf. Fusion 2025, 118, 102940. [Google Scholar] [CrossRef]
  23. Hannuna, S.; Camplani, M.; Hall, J.; Mirmehdi, M.; Damen, D.; Burghardt, T.; Paiement, A.; Tao, L. DS-KCF: A real-time tracker for RGB-D data. J. Real-Time Image Process. 2019, 16, 1439–1458. [Google Scholar] [CrossRef]
  24. Xiao, J.; Stolkin, R.; Gao, Y.; Leonardis, A. Robust fusion of color and depth data for RGB-D target tracking using adaptive range-invariant depth models and spatio-temporal consistency constraints. IEEE Trans. Cybern. 2018, 48, 2485–2499. [Google Scholar] [CrossRef] [PubMed]
  25. Jiang, M.; Deng, C.; Shan, J.; Wang, Y.; Jia, Y.; Sun, X. Hierarchical multi-modal fusion FCN with attention model for RGB-D tracking. Inf. Fusion 2019, 50, 1–8. [Google Scholar] [CrossRef]
  26. Kart, U.; Lukežič, A.; Kristan, M.; Kämäräinen, J.K.; Matas, J. Object tracking by reconstruction with view-specific discriminative correlation filters. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 1339–1348. [Google Scholar] [CrossRef]
  27. Gao, L.; Ke, Y.; Zhao, W.; Zhang, Y.; Jiang, Y.; He, G.; Li, Y. RGB-D visual object tracking with transformer-based multi-modal feature fusion. Knowl.-Based Syst. 2025, 322, 113531. [Google Scholar] [CrossRef]
  28. Zhang, J.; Yang, X.; Fu, Y.; Wei, X.; Yin, B.; Dong, B. Object tracking by jointly exploiting frame and event domain. In Proceedings of the 2021 IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 10–17 October 2021; pp. 13023–13032. [Google Scholar] [CrossRef]
  29. Yang, J.; Li, Z.; Zheng, F.; Leonardis, A.; Song, J. Prompting for multi-modal tracking. In Proceedings of the MM ’22: 30th ACM International Conference on Multimedia, Lisboa, Portugal, 10–14 October 2022; pp. 3492–3500. [Google Scholar] [CrossRef]
  30. Hong, L.; Yan, S.; Zhang, R.; Li, W.; Zhou, X.; Guo, P.; Jiang, K.; Chen, Y.; Li, J.; Chen, Z.; et al. OneTracker: Unifying visual object tracking with foundation models and efficient tuning. In Proceedings of the 2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 16–22 June 2024; pp. 19079–19091. [Google Scholar] [CrossRef]
  31. Shi, L.; Zhong, B.; Liang, Q.; Li, N.; Zhang, S.; Li, X. Explicit visual prompts for visual object tracking. In Proceedings of the Thirty-Eighth AAAI Conference on Artificial Intelligence (AAAI’24), Vancouver, BC, Canada, 20–27 February 2024; AAAI Press: Cambridge, MA, USA, 2024; Volume 38, pp. 4838–4846. [Google Scholar] [CrossRef]
  32. Ye, B.; Chang, H.; Ma, B.; Shan, S.; Chen, X. Joint feature learning and relation modeling for tracking: A one-stream framework. In Proceedings of the Computer Vision—ECCV 2022, Tel Aviv, Israel, 23–27 October 2022; Avidan, S., Brostow, G., Cissé, M., Farinella, G.M., Hassner, T., Eds.; Springer: Cham, Switzerland, 2022; pp. 341–357. [Google Scholar] [CrossRef]
  33. Li, C.; Xue, W.; Jia, Y.; Qu, Z.; Luo, B.; Tang, J.; Sun, D. LasHeR: A large-scale high-diversity benchmark for RGBT tracking. IEEE Trans. Image Process. 2022, 31, 392–404. [Google Scholar] [CrossRef] [PubMed]
  34. Loshchilov, I.; Hutter, F. SGDR: Stochastic gradient descent with warm restarts. arXiv 2017. [Google Scholar] [CrossRef]
  35. Glorot, X.; Bengio, Y. Understanding the difficulty of training deep feedforward neural networks. In Proceedings of the Thirteenth International Conference on Artificial Intelligence and Statistics. JMLR Workshop and Conference Proceedings, Chia Laguna Resort, Italy, 13–15 May 2010; pp. 249–256. [Google Scholar]
  36. Nam, H.; Han, B. Learning multi-domain convolutional neural networks for visual tracking. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 4293–4302. [Google Scholar] [CrossRef]
  37. Danelljan, M.; Van Gool, L.; Timofte, R. Probabilistic regression for visual tracking. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 7181–7190. [Google Scholar] [CrossRef]
  38. Dai, K.; Zhang, Y.; Wang, D.; Li, J.; Lu, H.; Yang, X. High-performance long-term tracking with meta-updater. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 6297–6306. [Google Scholar] [CrossRef]
  39. Chen, X.; Yan, B.; Zhu, J.; Wang, D.; Yang, X.; Lu, H. Transformer tracking. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 8122–8131. [Google Scholar] [CrossRef]
  40. Wang, H.; Liu, X.; Li, Y.; Sun, M.; Yuan, D.; Liu, J. Temporal adaptive RGBT tracking with modality prompt. In Proceedings of the AAAI Conference on Artificial Intelligence, Vancouver, BC, Canada, 20–27 February 2024; AAAI Press: Cambridge, MA, USA, 2024; Volume 38, pp. 5436–5444. [Google Scholar] [CrossRef]
  41. Ying, G.; Zhang, D.; Ou, Z.; Wang, X.; Zheng, Z. Temporal adaptive bidirectional bridging for RGB-D tracking. Pattern Recognit. 2025, 158, 111053. [Google Scholar] [CrossRef]
  42. Wu, Z.; Zheng, J.; Ren, X.; Vasluianu, F.A.; Ma, C.; Paudel, D.P.; Van Gool, L.; Timofte, R. Single-model and any-modality for video object tracking. In Proceedings of the 2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 16–22 June 2024; pp. 19156–19166. [Google Scholar] [CrossRef]
  43. Radford, A.; Kim, J.W.; Hallacy, C.; Ramesh, A.; Goh, G.; Agarwal, S.; Sastry, G.; Askell, A.; Mishkin, P.; Clark, J.; et al. Learning Transferable Visual Models From Natural Language Supervision. arXiv 2021. [Google Scholar] [CrossRef]
  44. Cheng, Z.; Chen, Q.; Zhang, J.; Fei, H.; Feng, X.; Che, W.; Li, M.; Qin, L. CoMT: A Novel Benchmark for Chain of Multi-modal Thought on Large Vision-Language Models. Proc. AAAI Conf. Artif. Intell. 2025, 39, 23678–23686. [Google Scholar] [CrossRef]
Figure 1. The framework of MPVT.
Figure 1. The framework of MPVT.
Applsci 15 07967 g001
Figure 2. The framework of dynamic adaptive prompt fusion.
Figure 2. The framework of dynamic adaptive prompt fusion.
Applsci 15 07967 g002
Figure 3. Precision in 15 degraded scenarios in the LasHeR dataset.
Figure 3. Precision in 15 degraded scenarios in the LasHeR dataset.
Applsci 15 07967 g003
Figure 4. Precision in 15 degraded scenarios in the VisEvent dataset.
Figure 4. Precision in 15 degraded scenarios in the VisEvent dataset.
Applsci 15 07967 g004
Figure 5. Sensitivity analysis of prompt depth across three datasets.
Figure 5. Sensitivity analysis of prompt depth across three datasets.
Applsci 15 07967 g005
Figure 6. Visualization of tracking in three multi-modal scenarios.
Figure 6. Visualization of tracking in three multi-modal scenarios.
Applsci 15 07967 g006
Table 1. The model parameter settings.
Table 1. The model parameter settings.
Parameter NameValue
Layers (N)40
Hidden dimensions (d)1536
Size of searching image ( H x = W x )378
Size of template image ( H z = W z )196
Feature size of searching region27
Feature size of template region14
Epochs60
Batch size16
Learning rateMaximal rate was 10 4 ; minimal rate was 10 6
Table 2. The results on LasHeR, DepthTrack, and VisEvent.
Table 2. The results on LasHeR, DepthTrack, and VisEvent.
Method R TP *LasHeRDepthTrackVisEvent
Sr (↑) Pr (↑) F 1 (↑) TRe (↑) Pr (↑) Sr (↑) Pr (↑)
TBSI [4]100%0.5060.638-----
APFNet [5]100%0.3620.500-----
FANet [17]100%0.3090.441-----
CAT [12]100%0.3140.450-----
MDNet [36]100%-----0.4260.661
PrDIMP50 [37]100%-----0.4530.644
LTMU [38]100%--0.4600.4170.5120.4590.655
OSTrack [32]100%0.4120.5150.5290.5220.5360.5340.695
TransT [39]100%0.3940.524---0.4740.650
DAL [8]100%--0.4290.3690.512--
TATrack [40]100%0.5590.702-----
TABBTrack [41]100%--0.6180.6150.622--
ProTrack [29]<1%0.4200.5380.5780.5730.5830.4710.632
ViPT [19]0.4%0.5250.6510.5940.5960.5920.5920.758
Un-Track [42]6.74%0.5360.667---0.5890.755
MPVT0.9%0.5780.7240.6280.6260.6310.6240.761
* R T P —ratio of trainable parameters.
Table 3. Ablation results on LasHeR, DepthTrack, and VisEvent.
Table 3. Ablation results on LasHeR, DepthTrack, and VisEvent.
Method R TP LasHerDepthTrackVisEvent
Sr (↑) Pr (↑) F 1 (↑) TRe (↑) Pr (↑) Sr (↑) Pr (↑)
MPVT0.9%0.5780.7240.6280.6260.6310.6240.761
MPVT-FC0.8%0.5620.7110.6130.6120.6140.6090.758
MPVT-PF100%0.5230.6540.5700.5690.5710.5940.754
MPVT-IE100%0.5170.6480.5560.5540.5570.5870.752
R T P —ratio of trainable parameters.
Table 4. Results of efficiency analysis.
Table 4. Results of efficiency analysis.
TypeTraining TimeGPU Memory UsageInference Speed
(Hours)↓ (GB)↓ (fps)↑
Full fine-tuning62.238.118
(removing prompt fusion module)
Prompt learning23.121.417
(enabling prompt fusion module)
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Xie, J.; Fu, Y.; Zhou, J.; He, T.; Wang, X.; Fang, Y.; Chen, D. MPVT: An Efficient Multi-Modal Prompt Vision Tracker for Visual Target Tracking. Appl. Sci. 2025, 15, 7967. https://doi.org/10.3390/app15147967

AMA Style

Xie J, Fu Y, Zhou J, He T, Wang X, Fang Y, Chen D. MPVT: An Efficient Multi-Modal Prompt Vision Tracker for Visual Target Tracking. Applied Sciences. 2025; 15(14):7967. https://doi.org/10.3390/app15147967

Chicago/Turabian Style

Xie, Jianyu, Yan Fu, Junlin Zhou, Tianxiang He, Xiaopeng Wang, Yuke Fang, and Duanbing Chen. 2025. "MPVT: An Efficient Multi-Modal Prompt Vision Tracker for Visual Target Tracking" Applied Sciences 15, no. 14: 7967. https://doi.org/10.3390/app15147967

APA Style

Xie, J., Fu, Y., Zhou, J., He, T., Wang, X., Fang, Y., & Chen, D. (2025). MPVT: An Efficient Multi-Modal Prompt Vision Tracker for Visual Target Tracking. Applied Sciences, 15(14), 7967. https://doi.org/10.3390/app15147967

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop