Next Article in Journal
A Possible Way for the Detection and Identification of Dangerous Substances in Ternary Mixtures Using THz Pulsed Spectroscopy
Previous Article in Journal
Energy-Efficient Spatial Query-Centric Geographic Routing Protocol in Wireless Sensor Networks
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Parallel Correlation Filters for Real-Time Visual Tracking

College of Information Science and Technology, Engineering Research Center of Digitized Textile & Fashion Technology, Ministry of Education, DongHua University, Shanghai 201620, China
*
Author to whom correspondence should be addressed.
Sensors 2019, 19(10), 2362; https://doi.org/10.3390/s19102362
Submission received: 24 April 2019 / Revised: 15 May 2019 / Accepted: 20 May 2019 / Published: 22 May 2019
(This article belongs to the Section Physical Sensors)

Abstract

:
Correlation filter-based methods have recently performed remarkably well in terms of accuracy and speed in the visual object tracking research field. However, most existing correlation filter-based methods are not robust to significant appearance changes in the target, especially when the target undergoes deformation, illumination variation, and rotation. In this paper, a novel parallel correlation filters (PCF) framework is proposed for real-time visual object tracking. Firstly, the proposed method constructs two parallel correlation filters, one for tracking the appearance changes in the target, and the other for tracking the translation of the target. Secondly, through weighted merging the response maps of these two parallel correlation filters, the proposed method accurately locates the center position of the target. Finally, in the training stage, a new reasonable distribution of the correlation output is proposed to replace the original Gaussian distribution to train more accurate correlation filters, which can prevent the model from drifting to achieve excellent tracking performance. The extensive qualitative and quantitative experiments on the common object tracking benchmarks OTB-2013 and OTB-2015 have demonstrated that the proposed PCF tracker outperforms most of the state-of-the-art trackers and achieves a high real-time tracking performance.

1. Introduction

Visual tracking plays a core role in computer vision for its wide applications including video surveillance, robotics, driver-less vehicles, intelligent interaction, and various automatic systems [1]. The goal of visual tracking is to track the trajectory of the target that is initialized only by the bounding box from the first frame among the video sequences [2]. During tracking, the appearance of the target changes randomly and unpredictably when the target undergoes deformation, illumination variation, and rotation. It is one of the core issues determining the tracking accuracy and robustness in visual tracking. Although significant progress has been achieved in recent decades, visual tracking is still a challenging problem due to these factors.
In general, visual trackers can be broadly classified into two categories, generative trackers [3,4,5,6,7,8,9,10,11,12], and discriminative trackers [13,14,15,16,17,18,19,20,21,22,23,24,25]. Generative trackers describe the target in the real world by the target representation method in computer vision and establish a target appearance model dynamically to find a candidate most similar to the target appearance model in the video sequence. Therefore, generative trackers can reflect the similarity of the same kind of target [3]. However, it tends to produce a significant number of false positives and the learning process is complicated. Discriminative trackers extract the discriminative features of the target and utilize the method of classification in machine learning to search for the region most similar to the target and locate the position of the target [13]. Discriminative trackers can find the optimal classification surface between different categories and reflect the difference between heterogeneous data. The discriminative model is more elegant and efficient than the generative model. Therefore, the discriminative model-based methods have been widely applied in the visual tracking field.
Recently, the discriminative correlation filter-based (DCF-based) methods [13,14,15,16,17,18,19,20] ignited the interest of scholars for visual tracking due to their high accuracy and speed. The standard DCF methods utilize the only given sample by the initial bounding box and many synthetic samples to learn a linear classifier or linear regressor online to predict the position of the target [14]. These synthetic samples can be easily generated by the circular shift windows in the frequency domain. The DCF technique is a computationally efficient process in the frequency domain transformed by fast Fourier transform (FFT) [26,27]. Therefore, the DCF methods achieve significant improvement in terms of accuracy and speed. Despite their success, it is unreasonable that the DCF methods utilize a fixed learning rate to learn the linear classifier. The smaller learning rate means that the trained DCF cannot track the changes in the target appearance. The larger learning rate means that the trained DCF may produce model drifting caused by the significant changes in the target appearance, especially when the target undergoes deformation, illumination variation, and rotation. Furthermore, the correlation output of the DCF methods follows a 2-D Gaussian distribution which is smooth at the response peak. It is can easily cause the model to drift.
This paper proposes a novel parallel correlation filters (PCF) visual tracker to overcome these issues described above. The proposed method constructs two parallel correlation filters with different learning rates. The correlation filter with the bigger learning rate is used to track the appearance changes in the target. The correlation filter with the smaller learning rate is used to track the translation of the target. Through weighting the response maps of the two parallel correlation filters, the position and scale of the target can be predicted accurately. In addition, the standard DCF assumes that the correlation response obeys the Gaussian distribution which has an unreasonable smooth peak. The triangle distribution is also not suitable for the DCFs which has a small slope at the hillside of the correlation response. Hence, a new reasonable distribution of correlation response that organically merges the two distributions is proposed to replace the original Gaussian distribution to prevent the model from drifting. In this paper, the implementation codes are all open source in the following Github web: https://github.com/YijYang/PCF.git.
In summary, the main contributions of this paper include:
  • A novel tracking framework named PCF that constructs two parallel correlation filters to simultaneously track the appearance and positional changes in the target;
  • A new reasonable distribution of correlation output is proposed to replace the original Gaussian distribution to prevent the model from drifting to achieve higher tracking accuracy and robustness.
  • Extensive qualitative and quantitative experiments on OTB-2013 [28] and OTB-2015 [29] benchmarks have demonstrated that the proposed PCF tracker performs better than most of the state-of-the-art trackers and simultaneously achieves a real-time tracking performance.

2. Related Works

Single visual object tracking is a quite common task in the field of computer vision. It has long been a focus area studied by many researchers. Plenty of visual trackers, including generative trackers, DCF-based trackers, and deep learning-based trackers, have been proposed in this field. In this section, the most representative and the most relevant works will be briefly discussed below.

2.1. Generative Trackers

The generative trackers utilize the features of the target extracted from the previous frame to dynamically learn the target’s appearance model, and then search for the region that best matches the prior appearance model among the candidate regions as the location of the target in the subsequent frame. Hence, the critical part to generative trackers is to search for the most similar candidate when the target has changed to some extent in a video frame. Up to now, many representative generative trackers [3,4,5,6,7,8,9,10,11,12] have been proposed, e.g., robust scale-adaptive mean-shift for tracking (ASMS) [3], incremental visual tracking (IVT) [4], fragment-based tracking (Frag) [5], multitask-based tracking (MTT) [6,7], distribution fields for tracking (DFT) [8] and distractor-aware tracker (DAT) [9]. ASMS is a real-time tracking algorithm implemented by minimizing the distance between two probability density histograms and it can track the scale of the target adaptively. However, it is easily distracted by similar objects in the surroundings. DFT builds a feature descriptor utilizing the distribution fields that smooth the objective function without breaking the information about the pixel value of the target. DAT trains the color histograms model of the target, distractors, and the background, and it uses the Bayesian function to compute the probability of each pixel belonging to the target, distractor, or background. However, these generative trackers do not take advantage of the background information and the performance of the generative models are limited by the dimensions of the target’s features. Therefore, the generative trackers could be easily distracted by background regions with similar appearances during tracking and show less discriminative power when the background is more complex.

2.2. DCF-Based Trackers

In contrast with generative trackers, DCF-based trackers effectively take advantage of the differences between the target and the background, and achieve better discriminative power and generative ability to separate the target from the background when the background is more complex. Recently, discriminative correlation filters (DCFs) have been widely applied in the visual tracking field in computer vision. Heads from signal processing, David S. Bolme and colleagues firstly introduced the correlation filter to the visual tracking fields [2]. They proposed the minimum output sum of squared error (MOSSE) tracker in the article ‘Visual Object Tracking using Adaptive Correlation Filters’. It produced significant performances, with tracking speed up to 700 frame rates. Thereafter, numerous improved algorithms [13,14,15,16,17,18,19,20] based on DCFs have been proposed with accurate and robust tracking performance by sacrificing the tracking speed, e.g., discriminative scale space tracking (DSST) [13], fast discriminative scale space tracking (FDSST) [13], spatially regularized discriminative correlation filters (SRDCF) [14], and efficient convolution operators with hand-crafted features (ECO-HC) [15] have demonstrated excellent performance on the popular benchmarks OTB-2013 and OTB-2015. DSST trains the translation and scale correlation filters using the handcrafted features extracted from an image sequence respectively. The improved FDSST utilized the principal component analysis (PCA) function to decrease the dimension of the extracted features to speed up the DSST. SRDCF adopts a spatial regularized component to penalize the correlation filters coefficients to solve the boundary effects. ECO-HC utilize a factorized convolution operator and a conservative update strategy to improve the speed and performance of the correlation filter. However, all these methods generate the training samples online and update the target appearance model with a fixed learning rate. Utilizing a fixed learning rate to update the model is unreasonable. DCF-based trackers cannot track the changes in the target appearance when the learning rate is too small. In contrast, trackers may produce model drifting when the target undergoes deformation, illumination variation, and rotation.

2.3. Deep Learning-Based Trackers

In several popular benchmarks [28,29], the deep convolutional neural networks (CNNs) have performed significantly well, which encouraged more recent works to either apply deep CNN features in the DCF algorithm [30,31,32,33,34] or design deep frameworks [35,36,37,38,39,40] for more accurate and robust visual tracking, e.g., learning multi-domain convolutional neural networks for visual tracking (MDNet) [33], multi-cue correlation filters for robust visual tracking (MCCT) [34], and convolutional residual learning for visual tracking (CREST) [35]. MDNet trains a multi-domain network composed of shared layers and multiple branches of domain-specific layers to track the target in different image sequences. MCCT constructs multiple-cue DCF experts to track the target independently achieving the strength of the different type of features. CREST reformulates DCFs as a one-layer convolutional neural network and exploits the residual learning method to achieve the benefit of end-to-end training and tracking. In contrast to hand-crafted features, learning DCF trackers that utilize deep CNN features remarkably increase their accuracy and robustness against the different degrees of change in the appearance model of the target. However, they suffer from high complexity and poor real-time performance. It is quite computationally expensive to extract deep CNN features with a high dimensionality or to implement deep architectures, resulting in poor real-time performance.

3. Proposed Approach

This section first provides a concise description of the standard discriminative correlation filter framework, then shows the implementation of the proposed PCF in detail. Finally, the algorithm framework of the PCF is described precisely in Algorithm 1.

3.1. Standard Discriminative Correlation Filter

The standard DCF [2] has been widely studied by many researchers due to its superior tracking accuracy and speed. The framework of the classical DCF is explicitly described in Figure 1. The DCF tracker trains a correlation filter model efficiently in the frequency domain by applying the machine-learning technique to distinguish the target from the background. It then updates the model online, exploiting the features extracted from the detected result in the current frame.
DCF learns the correlation filter h t l online by exploiting n target samples f 1 l , f 2 l , , f n l , which are efficiently generated by all cyclic shifts of the training sample. The objective function of the correlation filter h t l is equal to an L 2 error function, which can be expressed as Equation (1):
ε ( h t ) = k = 1 n l = 1 d h l f k l g k 2 + λ l = 1 d h l 2
where means the convolutional operator. k , l denote the number of training samples and the total dimension of the extracted features respectively. The expected correlation output g k is a 2-dimensional Gaussian distribution with the same size of h t l . The second term in this Equation is a regularization term which is utilized to prevent overfitting and λ ( λ 0 ) is a regularization parameter.
For computational efficiency, Equation (1) can be transformed into the frequency domain by fast Fourier transform (FFT). Then the objective function can be transformed as follows:
min ( H l ) ( k = 1 n l = 1 d H ¯ l F k l G k 2 + λ l = 1 d H l 2 )
where the capital letters denote the discrete Fourier transformations. H t l denotes the correlation filter in the frequency domain and the overbar of H ¯ t l represents complex conjugation of H t l . The convolution operation can be efficiently implemented by element-wise product in the frequency domain.
Therefore, through deriving Equation (2) and setting the derivative as zero, the final solutions are computed by Equation (5), which is implemented as follows:
A t l = G ¯ F t l
B t = k = 1 d F ¯ t k F t k + λ
H t l = A t l B t , l = 1 , 2 , , d
where t is the current step of the current video frame. In Equation (5), H t l denotes the trained correlation filter in the frequency domain and l is the dimension value of the filter, here the dimension value is equal to the dimension of the extracted features. For computational efficiency, the solution can be divided into two part as Equations (3) and (4), A t l represents the numerator of the filter, and B t represents the denominator of the filter. The overbar of F ¯ means the complex conjugation.
In order to adapt to the photometric and geometric variations of the target appearance, A t l , B t of the correlation filter is updated online by Equations (6) and (7) respectively:
A t l = ( 1 η 0 ) A t 1 l + η 0 G ¯ F t l
B t = ( 1 η 0 ) B t 1 + η 0 k = 1 d F ¯ t k F t k
where the scalar η 0 is a fixed learning rate.
For detecting the position of the target in the frame t , the features Z l are extracted from the region of the target pending detection. The responding correlation scores y t can be computed by Equation (8). Then the position of the maximum value of y t is regarded as the center position of the target area in the current frame.
y t = F 1 { l = 1 d A ¯ t 1 l Z l B t 1 + λ }
where F 1 denotes the inverse fast Fourier transform (IFFT).
In this paper, the DCF-based tracker ECO-HC [15] was chosen as the baseline tracker due to its excellent performance in terms of accuracy and robustness. The ECO-HC tracker has made several improvements on the basis of the standard DCF tracker. Firstly, it factorizes the convolution operation to reduce the model parameters. Secondly, it simplifies the generation of the training set by merging similar samples into a component to guarantee the diversity of samples. Thirdly, it employs the sparse update strategy to avoid the model drift problem where the update interval is set to 6. However, ECO-HC tracker has its own deficiencies. It is unreasonable to update the correlation model with a fixed learning rate when the sample appearance changes greatly due to deformation, illumination variation or rotation. Hence, ECO-HC tracker is difficult to track the target when the target appearance changes greatly. Furthermore, the Gaussian distribution with a smooth peak is utilized as the desired output of the correlation filter, which makes the result of the location inaccurate. In contrast to the ECO-HC, the proposed PCF tracker has been improved on this issue and obtained remarkable improvements on the benchmarks as shown in Figure 2.

3.2. Parallel Correlation Filters

The fundamental framework of the proposed PCF is explicitly described in Figure 3. In the initial frame, the PCF tracker trains two parallel correlation filters PCF1 and PCF2 online in a frequency domain utilizing the shared samples and the sharp correlation output. In the current frame, PCF1 and PCF2 are utilized to track the target respectively. Through weighting the response maps of PCF1 and PCF2, PCF tracker can detect the position of the target by applying the Newton method. Then it utilizes the Gaussian mixture model (GMM) [15] to generate a new sample set by adding a new sample or merging the two closest samples. As with the update strategy in the baseline tracker [15], it utilizes the new sample set to update the two parallel correlation filters with different learning rates every six frames.
PCF learns respectively two parallel correlation filters PCF1 h 1 t l and PCF2 h 2 t l by exploiting m target samples f 1 k l , f 2 k l , f m k l with different weights w k p t which are determined by the learning rate. The objective function of PCF1 can be expressed as Equation (9) and the objective function of PCF2 can be expressed as Equation (10):
ε ( h 1 t ) = k = 1 n l = 1 d h 1 l p = 1 m ( w 1 k p f k p l ) g s k 2 + λ l = 1 d ω h 1 l 2
ε ( h 2 t ) = k = 1 n l = 1 d h 2 l p = 1 m ( w 2 k p f k p l ) g s k 2 + λ l = 1 d ω h 2 l 2
where k , p denote the number of samples in the training set and the circular shift set respectively. The sharp output g s k is merged by the Gaussian distribution and the triangle distribution which will be described in detail in Section 3.3. The second term of these two Equations is a regularized term and ω denotes the spatial regularization parameter.
Equations (9) and (10) can be transformed into the frequency domain efficiently by FFT. Due to the regularization parameter, ω breaks the closed solution of the objective functions, the solutions of H 1 t 1 l and H 2 t 1 l are obtained by the conjugation gradient (CG) iterative method [15]. Then for detecting the position of the target in the frame t , the shared features Z t , p o s l are extracted from the region of the target pending detection. The center position of the region is determined by the previously detected position. The corresponding correlation scores y t , p o s can be computed by Equation (11). Then the position of the target in the current frame is optimized by the Newton iterative method.
y t , p o s = F 1 { l = 1 d ( α H 1 t 1 l Z t , p o s l + ( 1 α ) H 2 t 1 l Z t , p o s l ) }
where α denotes the fusion factor. H 1 t 1 l and H 2 t 1 l represent two parallel correlation filters in the frequency domain.
After detection, the new sample is extracted from the tracking result. Then the GMM is utilized to compute the similarities between the new sample and the components of the training set. After that, a new training set is generated by adding the new sample or merging two closest samples. The weights w k p t of the trained samples are updated by Equation (12).
w k p t = ( 1 η ) w k p t 1
where t denotes the t-th frame and η is the learning rate. Different learning rates mean different weights of samples.
For further detecting the scale of the target [13] in the frame t , the standard DCF is utilized to extract scale features Z t , s c a l e l with different scale factors and to compute the scale correlation scores y t , s c a l e in Equation (13). Then the scale factor of the maximum value of y t , s c a l e is utilized to compute the scale of the target in the current frame.
y t , s c a l e = F 1 { l = 1 d A ¯ t 1 , s c a l e l Z t , s c a l e l B t 1 , s c a l e + λ }
where F 1 represents IFFT. A ¯ t 1 , s c a l e l , B t 1 , s c a l e denote the numerator and the denominator of the scale correlation filter in the previous frame receptively.

3.3. Reasonable Distribution of the Correlation Response

The Gaussian distribution, the triangle distribution, and the merged distribution are described clearly in Figure 4a–c respectively. The desired output of the standard DCF methods follows a 2-D Gaussian distribution with a smooth peak. Due to the DCF training, the model using the synthetic samples generated by circular shift windows, the peak of the desired output should be sharp to avoid model drifting. The 2-D Gaussian probability density g ( x , y ) is expressed in Equation (14).
g ( x , y ) = 1 2 π σ 2 e ( x 2 + y 2 ) 2 σ 2 , ( w 2 x w 2 , h 2 y h 2 )
where σ represents the standard deviation and it is set to 1/16 in this paper. x , y denotes the coordinates for the figure pixels. w , h denotes the width and height of the figure, and ( w 2 , h 2 ) is regarded as the original point.
Different from the Gaussian distribution, the triangle distribution shown in Figure 4b has a sharp peak. However, it has a small slope at the hillside of the distribution. Due to the synthetic samples are generated by circular shift windows, the positions below the hillside of the distribution are regarded as the label of negative samples. The values of these positions should set nearly to zero. The 2-D triangle probability density s ( x , y ) is expressed in Equation (15).
s ( x , y ) = ( 1 2 w | x | ) ( 1 2 h | y | ) , ( w 2 x w 2 , h 2 y h 2 )
where ( w 2 , h 2 ) is regarded as the original point.
Both Gaussian distribution and triangle distribution are not suitable for DCF. Hence, a new sharp distribution of the desired output is proposed to replace the original Gaussian distribution to achieve higher tracking accuracy and robustness. The 2-D merged probability density m ( x , y ) is expressed in Equation (16). The sharp distribution organically merges Gaussian distribution and triangle distribution by multiplication operators. It simultaneously combines the merits of these two distributions, with a sharp peak and a large slope at the hillside. Extensive experiments on benchmarks have demonstrated the effectiveness of the sharp distribution.
m ( x , y ) = 1 2 π σ 2 ( 1 2 w | x | ) ( 1 2 h | y | ) e ( x 2 + y 2 ) 2 σ 2 , ( w 2 x w 2 , h 2 y h 2 )
where ( w 2 , h 2 ) is regarded as the original point.

3.4. The Brief Outline of the PCF Tracking Algorithm

The proposed PCF tracking algorithm is briefly described in Algorithm 1. The PCF tracker first detects the position of the target, and then the scale correlation filter is utilized to refine the scale of the target. There are some differences between the scale and the position correlation filters. For the position filters, the PCF tracker extracts the position features from the first frame with a bounding box. It then uses the extracted features initializes two parallel correlation filters PCF1 H 1 1 l and PCF2 H 2 1 l online in the frequency domain with a sharp correlation output. From the second frame to the end frame of the sequence, PCF1 and PCF2 are utilized to track the target respectively. Through weighting the response maps of these two parallel correlation filters, the PCF tracker can detect the center position P t of the target by applying the NM. Then it utilizes the GMM to generate a new sample set. It also utilizes the features extracted from the new sample set to update the two parallel correlation filters H 1 t l , H 2 t l with different learning rates η t 1 , η t 2 every six frames. For the scale filter, the PCF tracker extracts the scale features from the first frame with different scale factors. It then uses the extracted features initializes the scale correlation filters A 1 , s c a l e , B 1 , s c a l e . In the subsequent frame of the sequence, similar to the position filters, the scale filter utilizes the detected result S t to update the filter A t , s c a l e , B t , s c a l e by the learning rate η 0 every frame.
Algorithm 1. PCF Tracking Algorithm
Input:
1: Image I t .
2: Detected target position P t 1 and scale S t 1 in the previous frame.
Output:
1: Detected target position P t and scale S t in the current frame.
Loop:
1: Initialize the PCF model H 1 1 l , H 2 1 l and scale model A 1 , s c a l e , B 1 , s c a l e in the first frame by Equations (9), (10) and (4), (5).
2: for t [ 2 , t f ] do.
3:  Position detection:
4:   Extract position features Z t , p o s from I t at P t 1 and S t 1 by a search region.
5:   Compute two parallel correlation scores y 1 t , p o s , y 2 t , p o s .
6:   Merge the two correlation scores to y t , p o s by Equation (11).
7:   Set P t to the target position by Newton iterative method.
8:  Scale detection:
9:   Extract scale feature Z t , s c a l e from I t at P t 1 and S t 1 by a search region.
10:    Compute correlation scores y t , s c a l e by Equation (13).
11:    Set S t to the target scale that maximizes y t , s c a l e .
12:   Model update every six frames:
13:    Extract new sample features F t , p o s and F t , s c a l e from I t at P t and S t .
14:    Generate new training set by the GMM method.
15:    Update the PCF model H 1 t l , H 2 t l by the learning rate η t 1 , η t 2 .
16:    Update the scale model A t , s c a l e , B t , s c a l e by the learning rate η 0 .
17:   Return P t , S t .
18: end for.

4. Experiments

In this section, the implementation materials and parameter settings are described in detail. Comprehensive experiments are then performed on two benchmarks OTB-2013 [28] and OTB-2015 [29] to validate the effectiveness of the PCF tracker. Finally, the tracking failure cases of the proposed PCF tracker are analyzed briefly, and the future improvement works are further presented on this basis. The results of these experiments have demonstrated that the proposed PCF tracker performs better than most of the state-of-the-art methods.

4.1. Implementation Details

All experiments are performed on the same desktop (equipping with INTEL i5-4590 CPU with 8G RAM). The PCF tracker presented in this paper is implemented in MATLAB R2016a. The relevant parameters of the PCF tracker are briefly described in Table 1. For parameter adjusting, the extensive parameter setting experiments are conducted on OTB-2013 benchmark. Figure 5 shows that the PCF tracker achieved the best OPE accuracy at the point (0.9, 0.5). Therefore, the learning rate η t 2 of PCF2 is set to 0.5 in Equation (12), and the merging factor α of the two response maps is set to 0.9 in Equation (11). Following the baseline tracker [15], the learning rate η t 1 of PCF1 is set to 0.009 in Equation (12). For the generative model of the training set mentioned in Section 3.2, the initial weight w k p 0 of the training sample is set to 1.0 and the number of the target samples m is set to 30. For the scale filter, the relevant parameters are set to the same values as the baseline tracker proposed in [19]. In order to make a fair comparison with the state-of-the-art trackers, the same parameters were used for all experiments on the benchmark data sets.

4.2. Ablation Experiments

To evaluate the effectiveness of progressively integrating the strategies proposed in this paper, the ablation experiments are conducted on OTB-2013 benchmark and the proposed PCF tracker is compared with the baseline tracker introduced in Section 3.1, the PCF with weighting fusion correlation response maps (PCF_WF tracker) proposed in Section 3.2 and the PCF with sharp output response (PCF_SR tracker) described in Section 3.3. The performance of these trackers is evaluated both on precision at different center location error thresholds and success rate at different overlap thresholds.
Figure 6 explicitly shows the comparison results on OTB-2013 benchmark in terms of the precision plots (PP) and success plots (SP) of one pass evaluation (OPE). The proposed strategies of weighting fusion and sharp output response in this paper either obtain significant improvement compared to the baseline tracker. In terms of the OPE of location error threshold (LET) at 20 pixels, the PCF_WR tracker and PCF_SR tracker achieve a gain of 3.3% and 3.1% compared to the baseline tracker respectively. From the aspect of the OPE of areas under the curve (AUC), the PCF_WR tracker and the PCF_SP tracker obtain a gain of 1.9% and 1.7% contrasted to the baseline tracker respectively. It indicates that both the two strategies are effective. Overall, the proposed PCF tracker integrating these two strategies performs the best performance in terms of precision and success rate. Concretely, the PCF tracker achieves 89.2%, 67.5% in the OPE of LET at 20 pixels and AUC, and compared to the baseline tracker, the PCF tracker achieves a remarkable gain of 4.3% and 3.2%.
Table 2 and Table 3 demonstrate separately the PP and SP of the OPE for the compared trackers in detail on eleven different attributes: scale variation (SV), illumination variation (IV), out-of-plane rotation (OPR), occlusion (OCC), background cluttered (BC), deformation (DEF), motion blur (MB), fast motion (FM), in-plane rotation (IPR), out-of-view (OV), and low resolution (LR). Obviously, the PCF_WF tracker and PCF_SR tracker all achieve various degrees of progress versus the baseline tracker, and perform well in real-time (average speed 44 FPS and 47 FPS respectively) in the benchmark. In Table 2, comparing PCF_WF tracker with PCF_SR tracker, the former performs better than the latter in terms of IV, OPR, BC, DEF, and IPR. However, from the aspects of SV, OCC, MB, FM, OV, and LR, the latter outperforms the former. This indicates that the strategy of weighting fusion is more robust to significant appearance changes in the target, especially when the target undergoes deformation, illumination variation, and rotation, while the scheme of sharp response is more accurate to the problem of model drifting caused by distractors or blur. In general, Table 2 and Table 3 indicate that the proposed PCF tracker acquires the best or the second results in terms of PP and SP of the OPE on all eleven attributes. Simultaneously, the PCF tracker runs an average speed of 42 FPS on all sequences of OTB-2013.

4.3. Experiments on OTB-2013

OTB-2013 is a common benchmark with 50 sequences that are divided into 11 different attributes: SV, IV, OPR, OCC, BC, DEF, MB, FM, IPR, OV, LR. On this challenge benchmark, the proposed PCF tracker is compared with 18 state-of-the-art trackers from the works: tracking-learning-detection (TLD) [2], distribution fields for tracking (DFT) [8], discriminative scale space tracking (DSST) [13], fast discriminative scale space tracking (FDSST) [13], spatially regularized discriminative correlation filters (SRDCF) [14], sum of template and pixel-wise learners (Staple) [16], compressive tracking (CT) [17], long-term correlation tracking (LCT) [18], locally orderless tracking (LOT) [19], least soft-threshold squares tracking (LSS) [21], visual tracking with online multiple instance learning (MIL) [22], scale adaptive kernel correlation filter tracker with feature integration (SAMF) [23], exploiting the circulant structure of tracking-by-detection with kernels (CSK) [24], high-speed tracking with kernelized correlation filters (KCF) [25], adaptive decontamination of the training set: A unified formulation for discriminative visual tracking (SRDCFdecon) [31], convolutional features for correlation filter-based visual tracking (DeepSRDCF) [32], Fully-convolutional Siamese networks for object tracking (SiamFC_3s) [36], object tracking via dual linear structured SVM and explicit feature map (DLSSVM) [37]. Only the ranks for the top 10 trackers are reported.
Figure 7 clearly illustrates the PP and SP of the OPE on three different attributes: DEF with 19 sequences, OPR with 39 sequences and IV with 25 sequences. As is shown in Figure 7, the proposed PCF achieves the best results among the top 10 trackers in these attributes. Specifically, the PCF tracker obtains 91.3% and 69.6%; 88.4% and 65.9%, and 82.1% and 62.3% on attributes DEF, OPR, and IV, respectively. It indicates that the PCF tracker is more accurate and robust to the significant changes in the target appearance compared with the other top nine trackers. Furthermore, the proposed PCF tracker is very effective for handling the challenges of deformation, illumination variation, and out-of-plane rotation.
The overall results on OTB-2013 benchmark among the top ten trackers are illustrated in Figure 8. Contrasted to the second tracker deepSRDCF-based on deep features, the proposed PCF tracker obtains the best ranks of 89.2%, 67.5% on the PP and SP of the OPE, and achieves a visibly gain of 3.9%, 7.5% in the PP and SP of the OPE respectively. Among the compared trackers employing handcrafted features, deep features, or combining these two features, the proposed PCF tracker achieves the best ranks and simultaneously runs a real-time speed of 41 FPS on a CPU. Furthermore, Table 4 and Table 5 explicitly shows the SP and PP of the OPE for the top 10 trackers on 11 challenge attributes respectively. As is demonstrated in Table 4 and Table 5, the PCF tracker outperforms the other top nine trackers on 10 out of 11 attributes of the OPE and obtains the best average precision (AP) and areas under the curve (AUC). It validates that the proposed sharp response strategy can effectively prevent the model from drifting and the proposed weighting fusion strategy is also very efficient to track the changes in the target appearance. Both strategies bring remarkable improvements in terms of accuracy and robustness. Besides, qualitative experiments are conducted on this benchmark and the results are reported in Figure 9. It further indicates that the proposed PCF tracker can accurately track the target with deformation, rotation, and illumination variation.

4.4. Experiments on OTB-2015

OTB-2015 is a more challenging benchmark than OTB-2013, including 100 videos with 11 different attributes: SV, IV, OPR, OCC, BC, DEF, MB, FM, IPR, OV, LR. The proposed PCF tracker is evaluated with 18 state-of-the-art trackers on this benchmark from the works: TLD [2], Incremental learning for robust visual tracking (IVT) [4], DFT [8], DSST [13], FDSST [13], SRDCF [14], Staple [16], CT [17], LCT [18], LOT [19], LSS [21], MIL [22], SAMF [23], CSK [24], KCF [25], SRDCFdecon [31], DeepSRDCF [32], and DLSSVM [37]. Only the ranks for the top 10 trackers are reported.
Figure 10 reports the PP and SP of the OPE determined by three challenges attributes: OPR with 63 videos, IV with 38 videos, and DEF with 44 videos. As is clearly illustrated in Figure 10, the proposed PCF provides the best results among the top 10 trackers in all three attributes. More particularly, the PCF tracker obtains 85.1% and 62.6%; 82.5% and 63.6%, and 83.0% and 62.5% on attributes OPR, IV, and DEF respectively. It again validates the effectiveness of the proposed PCF tracker for handling the issues of the significant changes in the target appearance.
Figure 11 shows the overall results for the top ten trackers on OTB-2015 benchmark. Among these compared trackers, the proposed PCF tracker obtains the best ranks in PP and SP of the OPE including the AP scores of 86.3% and the AUC scores of 64.7%. In addition, attributes-based evaluations are conducted on this benchmark. The results of this experiment are reported in Table 6 and Table 7. Specifically, the proposed PCF tracker achieves the top scores in terms of PP on 9 out of 11 attributes as shown in Table 7, and obtains the best ranks in terms of SP on all 11 attributes as demonstrated in Table 7. As is demonstrated explicitly in Table 6 and Table 7, the proposed PCF tracker obtains the best AP and AUC scores compared to the other top nine trackers, at the same time, the PCF tracker achieves a real-time performance running about 41 FPS on a CPU. It indicates the superior tracking performance for all 11 attributes. In general, the proposed tracker achieves a substantial improvement of the other top nine trackers in terms of accuracy, robustness and real-time performance. Furthermore, the qualitative experiments are also conducted in all videos of OTB-2015 benchmark. The results for the four representative videos are illustrated in Figure 12. Among the compared trackers, the proposed PCF tracker significantly outperforms the other top nine trackers in terms of location and scale estimation. It again validates that the effectiveness of the proposed PCF tracker when the target undergoes the situations of deformation, rotation, and illumination variation.

4.5. Analysis of the Failure Cases

Figure 13 shows the failure cases of the proposed PCF tracker. In the first row, the target underwent complete occlusions for long spans of time which caused the PCF tracker to drift off the target. While the LCT tracker can still track the target because of the re-detection strategy [18]. In the second row, the target underwent heavy occlusion with background clutters which also resulted in the PCF tracker failure to track the target. However, the trackers with deep features (e.g., deepSRDCF and SRDCFdecon) have high robustness in this situation. Furthermore, in the field of multi-object tracking, the top-down Bayesian formulation proposed in the work [41] can also solve the problems of occlusion and background clutters effectively. Hence, in the future works, the works [14,18,31,41] will be analyzed in detail and the strategies of these works will be merged into the proposed PCF tracker to address these issues.

5. Conclusions

In this paper, a novel PCF framework was proposed to address the issues of target appearance changes and model drifting. The proposed method constructed two parallel correlation filters with different learning rate. The correlation filter with the bigger learning rate was proposed to track the appearance changes in the target. The correlation filter with the smaller learning rate was applied to track the location of the target. Through weighting the response maps of the two parallel correlation filters, the position and scale of the target can be estimated accurately. Furthermore, a new reasonable distribution of correlation response that organically merges the Gaussian distribution and the triangle distribution was proposed to replace the original Gaussian distribution to prevent the model from drifting to achieve higher tracking accuracy and robustness. Extensive qualitative and quantitative evaluations on serval common benchmarks have demonstrated the competitive accuracy, robustness, and the superior tracking speed performance of the proposed PCF tracker compared to the state-of-the-art trackers. After analyzing the failure cases of the proposed PCF tracker, a new re-detection strategy will be studied in detail to improve the disadvantages of heavy or complete occlusions in visual tracking.

Author Contributions

Y.Y. and Y.Z. conceived the main idea, designed the main algorithm, and wrote the manuscript. Y.Y. designed the main experiments under the supervision of Y.Z., D.L., and Z.W. The experimental results were analyzed by Y.Y., Y.Z., D.L., and Z.W. provided suggestions for the proposed algorithm.

Funding

This research was funded by [the Fundamental Research Funding for the Central Universities of Ministry of Education of China] grant number [18D110408], [the National Natural Science Foundation of China (NSFC)] grant number [18K10454], and [the Special Project Funding for the Shanghai Municipal Commission of Economy and Information Civil-Military Inosculation Project “Big Data Management System of UAVs”] grant number [JMRH-2018-1042]. The APC was funded by [the Fundamental Research Funding for the Central Universities of Ministry of Education of China].

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Kalal, Z.; Mikolajczyk, K.; Matas, J. Tracking-Learning-Detection. IEEE Trans. Pattern Anal. Mach. Intell. 2012, 34, 1409–1422. [Google Scholar] [CrossRef] [PubMed]
  2. Bolme, D.S.; Beveridge, J.R.; Draper, B.A.; Lui, Y.M. Visual object tracking using adaptive correlation filters. In Proceedings of the 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR), San Francisco, CA, USA, 13–18 June 2010; Volume 119, pp. 2544–2550. [Google Scholar]
  3. Vojir, T.; Noskova, J.; Matas, J. Robust Scale-Adaptive Mean-Shift for Tracking. Pattern Recognit. Lett. 2014, 49, 250–258. [Google Scholar] [CrossRef]
  4. Ross, D.A.; Lim, J.; Lin, R.S.; Yang, M.H. Incremental learning for robust visual tracking. Int. J. Comput. Vis. 2008, 77, 125–141. [Google Scholar] [CrossRef]
  5. Adam, A.; Rivlin, E.; Shimshoni, I. Robust fragments-based tracking using the integral histogram. In Proceedings of the 2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR), New York, NY, USA, 17–22 June 2006; pp. 798–805. [Google Scholar]
  6. Zhang, T.; Ghanem, B.; Liu, S.; Ahuja, N. Robust visual tracking via multi-task sparse learning. In Proceedings of the 2012 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Providence, RI, USA, 16–21 June 2012; pp. 2042–2049. [Google Scholar]
  7. Zhang, T.; Ghanem, B.; Liu, S.; Ahuja, N. Robust visual tracking via structured multi-task sparse learning. Int. J. Comput. Vis. 2013, 101, 367–383. [Google Scholar] [CrossRef]
  8. Sevilla-Lara, L. Distribution fields for tracking. In Proceedings of the 2012 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Providence, RI, USA, 16–21 June 2012; Volume 2012, pp. 1910–1917. [Google Scholar]
  9. Possegger, H.; Mauthner, T.; Bischof, H. In defense of color-based model-free tracking. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; Volume 2015, pp. 2113–2120. [Google Scholar]
  10. Danelljan, M.; Khan, F.S.; Felsberg, M.; Weijer, J.V.D. Adaptive Color Attributes for Real-Time Visual Tracking. In Proceedings of the 2012 IEEE Conference on Computer Vision and Pattern Recognition, Providence (CVPR), Columbus, OH, USA, 23–28 June 2014; Volume 2014, pp. 1090–1097. [Google Scholar]
  11. He, S.; Yang, Q.; Lau, R.W.H.; Wang, J.; Yang, M.H. Visual Tracking via Locality Sensitive Histograms. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Portland, OR, USA, 23–28 June 2013; Volume 2013, pp. 2427–2434. [Google Scholar]
  12. Lu, H.; Jia, X.; Yang, M.H. Visual tracking via adaptive structural local sparse appearance model. In Proceedings of the IEEE Computer Vision and Pattern Recognition (CVPR), Providence, RI, USA, 16–21 June 2012; Volume 2012, pp. 1822–1829. [Google Scholar]
  13. Danelljan, M.; Häger, G.; Khan, F.S.; Felsberg, M. Discriminative Scale Space Tracking. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 1561–1575. [Google Scholar] [CrossRef] [PubMed]
  14. Danelljan, M.; Hager, G.; Khan, F.S.; Felsberg, M. Learning Spatially Regularized Correlation Filters for Visual Tracking. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, 7–13 December 2015; Volume 2015, pp. 4310–4318. [Google Scholar]
  15. Danelljan, M.; Bhat, G.; Khan, F.S.; Felsberg, M. ECO: Efficient Convolution Operators for Tracking. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; Volume 2017, pp. 6931–6939. [Google Scholar]
  16. Bertinetto, L.; Valmadre, J.; Golodetz, S.; Miksik, O.; Torr, P.H. Staple: Complementary Learners for Real-Time Tracking. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; Volume 2016, pp. 1401–1409. [Google Scholar]
  17. Zhang, K.; Zhang, L.; Yang, M.H. Real-Time Compressive Tracking. In Proceedings of the IEEE European Conference on Computer Vision (ECCV), Florence, Italy, 7–13 October 2012; Volume 2012, pp. 864–877. [Google Scholar]
  18. Ma, C.; Yang, X.; Zhang, C.; Yang, M.H. Long-term correlation tracking. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; Volume 2015, pp. 5388–5396. [Google Scholar]
  19. Oron, S.; Bar-Hillel, A.; Levi, D.; Avidan, S. Locally Orderless Tracking. In Proceedings of the 2012 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Providence, RI, USA, 16–21 June 2012; Volume 2012, pp. 1940–1947. [Google Scholar]
  20. Lukezic, A.; Vojir, T.; Zajc, L.C.; Matas, J.; Kristan, M. Discriminative Correlation Filter Tracker with Channel and Spatial Reliability. Int. J. Comput. Vis. 2018, 126, 671–688. [Google Scholar] [CrossRef] [Green Version]
  21. Wang, D.; Lu, H.; Yang, M.H. Least Soft-Threshold Squares Tracking. In Proceedings of the IEEE Computer Vision and Pattern Recognition (CVPR), Portland, OR, USA, 23–28 June 2013; Volume 2013, pp. 2371–2378. [Google Scholar]
  22. Babenko, B.; Yang, M.H.; Belongie, S. Visual tracking with online multiple instance learning. In Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Miami Beach, FL, USA, 20–25 June 2009; pp. 983–990. [Google Scholar]
  23. Li, Y.; Zhu, J. A Scale Adaptive Kernel Correlation Filter Tracker with Feature Integration. In Proceedings of the IEEE European Conference on Computer Vision Workshops (ECCVW), Zurich, Switzerland, 6–12 September 2014; Volume 8926, pp. 254–265. [Google Scholar]
  24. Rui, C.; Martins, P.; Batista, J. Exploiting the circulant structure of tracking-by-detection with kernels. In Proceedings of the IEEE European Conference on Computer Vision (ECCV), Florence, Italy, 7–13 October 2012; Volume 2012, pp. 702–715. [Google Scholar]
  25. Joao, F.H.; Rui, C.; Pedro, M.; Jorge, B. High-Speed Tracking with Kernelized Correlation Filters. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 37, 583–596. [Google Scholar]
  26. Yang, R.; Wei, Z. Real-Time Visual Tracking through Fusion Features. Sensors 2016, 16, 949. [Google Scholar] [CrossRef]
  27. Shi, G.; Xu, T.; Guo, J.; Luo, J.; Li, Y. Consistently Sampled Correlation Filters with Space Anisotropic Regularization for Visual Tracking. Sensors 2017, 17, 2889. [Google Scholar] [CrossRef]
  28. Wu, Y.; Lim, J.; Yang, M.H. Online Object Tracking: A Benchmark. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Portland, OR, USA, 23–28 June 2013; Volume 9, pp. 2411–2418. [Google Scholar]
  29. Wu, Y.; Lim, J.; Yang, M.H. Online Object Tracking: A Benchmark. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 37, 1834–1848. [Google Scholar] [CrossRef] [PubMed]
  30. Danelljan, M.; Robinson, A.; Khan, F.S.; Felsberg, M. Beyond Correlation Filters: Learning Continuous Convolution Operators for Visual Tracking. In Proceedings of the IEEE European Conference on Computer Vision (ECCV), Amsterdam, The Netherlands, 8–10 October 2016; Volume 2016, pp. 472–488. [Google Scholar]
  31. Danelljan, M.; Hager, G.; Khan, F.S.; Felsberg, M. Adaptive decontamination of the training set: A unified formulation for discriminative visual tracking. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; Volume 2016, pp. 1430–1438. [Google Scholar]
  32. Danelljan, M.; Hager, G.; Khan, F.S.; Felsberg, M. Convolutional features for correlation filter based visual tracking. In Proceedings of the IEEE International Conference on Computer Vision Workshop (ICCVW), Santiago, Chile, 7–13 December 2015; Volume 2015, pp. 621–629. [Google Scholar]
  33. Nam, H.; Han, B. Learning Multi-Domain Convolutional Neural Networks for Visual Tracking. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; Volume 2015, pp. 4293–4302. [Google Scholar]
  34. Wang, N.; Zhou, W.; Tian, Q.; Hong, R.; Wang, M.; Li, H. Multi-Cue Correlation Filters for Robust Visual Tracking. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–23 June 2018; Volume 2018, pp. 4844–4853. [Google Scholar]
  35. Song, Y.; Ma, C.; Gong, L.; Zhang, J.; Lau, R.; Yang, M.H. Crest: Convolutional residual learning for visual tracking. In Proceedings of the 2017 IEEE International Conference on Computer Vision Workshop (ICCV), Venice, Italy, 22–29 October 2017; Volume 2017, pp. 2574–2583. [Google Scholar]
  36. Bertinetto, L.; Valmadre, J.; Henriques, J.F.; Vedaldi, A.; Torr, P.H.S. Fully-Convolutional Siamese Networks for Object Tracking. In Proceedings of the IEEE European Conference on Computer Vision (ECCV), Amsterdam, The Netherlands, 8–10 October 2016; Volume 2016, pp. 850–865. [Google Scholar]
  37. Ning, J.; Yang, J.; Jiang, S.; Zhang, L.; Yang, M.-H. Object Tracking via Dual Linear Structured SVM and Explicit Feature Map. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; Volume 2016, pp. 4266–4274. [Google Scholar]
  38. Fan, H.; Ling, H. SANet: Structure-Aware Network for Visual Tracking. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Honolulu, HI, USA, 21–26 July 2017; Volume 2017, pp. 2217–2224. [Google Scholar]
  39. Zhang, T.; Xu, C.; Yang, M.H. Multi-task Correlation Particle Filter for Robust Object Tracking. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; Volume 2017, pp. 4819–4827. [Google Scholar]
  40. Pu, S.; Song, Y.; Ma, C.; Zhang, H.; Yang, M.-H. Deep Attentive Tracking via Reciprocative Learning. In Proceedings of the 32nd Conference on Neural Information Processing Systems (NIPS), Montreal, QC, Canada, 2–8 December 2018; Volume 1810, pp. 3851–3866. [Google Scholar]
  41. Kim, D.Y.; Vo, B.N.; Vo, B.T.; Jeon, M. A labeled random finite set online multi-object tracker for video data. Pattern Recognit. 2019, 90, 377–389. [Google Scholar] [CrossRef]
Figure 1. Illustration of the standard DCF framework. Firstly, the correlation filter is trained by the first frame with an initial bounding box and the desired output. Secondly, it extracts the features from the current frame and multiplies a cosine window to emphasis on the center region. The features are then transformed into the frequency domain by utilizing the fast Fourier transformation (FFT). Thirdly, the response map is obtained by multiplying the correlation filter and the extracted features. The response map is transformed into the time domain by applying the inverse fast Fourier transformation (IFFT). Finally, the position of the maximum value of the response map is regarded as the center position of the target in the current frame. The new features are then extracted from the detected result to train and update the correlation filter.
Figure 1. Illustration of the standard DCF framework. Firstly, the correlation filter is trained by the first frame with an initial bounding box and the desired output. Secondly, it extracts the features from the current frame and multiplies a cosine window to emphasis on the center region. The features are then transformed into the frequency domain by utilizing the fast Fourier transformation (FFT). Thirdly, the response map is obtained by multiplying the correlation filter and the extracted features. The response map is transformed into the time domain by applying the inverse fast Fourier transformation (IFFT). Finally, the position of the maximum value of the response map is regarded as the center position of the target in the current frame. The new features are then extracted from the detected result to train and update the correlation filter.
Sensors 19 02362 g001
Figure 2. The qualitative comparisons of the proposed parallel correlation filters (PCF) tracker with the baseline tracker efficient convolution operators with hand-crafted features (ECO-HC) on five challenging sequences of the benchmarks. The results are marked in different colors. On all these five cases, PCF tracker performs better center position precision and overlap precision than the baseline tracker. PCF tracker successfully tracks the target when the target undergoes significant rotation, illumination variation, and deformation.
Figure 2. The qualitative comparisons of the proposed parallel correlation filters (PCF) tracker with the baseline tracker efficient convolution operators with hand-crafted features (ECO-HC) on five challenging sequences of the benchmarks. The results are marked in different colors. On all these five cases, PCF tracker performs better center position precision and overlap precision than the baseline tracker. PCF tracker successfully tracks the target when the target undergoes significant rotation, illumination variation, and deformation.
Sensors 19 02362 g002
Figure 3. The flow chart of the proposed PCF framework. In the initial frame, the PCF tracker trains two parallel correlation filters PCF1 and PCF2 online in the frequency domain utilizing the shared samples and the sharp correlation output. In the current frame, PCF1 and PCF2 are utilized to track the target respectively. Through weighting the response maps of PCF1 and PCF2, the PCF tracker accurately detects the position of the target by applying the Newton method (NM). It then utilizes the Gaussian mixture model (GMM) to generate a new shared sample set by adding a new sample or merging the two closest samples. It utilizes the new sample set to update the two parallel correlation filters with different learning rates every six frames.
Figure 3. The flow chart of the proposed PCF framework. In the initial frame, the PCF tracker trains two parallel correlation filters PCF1 and PCF2 online in the frequency domain utilizing the shared samples and the sharp correlation output. In the current frame, PCF1 and PCF2 are utilized to track the target respectively. Through weighting the response maps of PCF1 and PCF2, the PCF tracker accurately detects the position of the target by applying the Newton method (NM). It then utilizes the Gaussian mixture model (GMM) to generate a new shared sample set by adding a new sample or merging the two closest samples. It utilizes the new sample set to update the two parallel correlation filters with different learning rates every six frames.
Sensors 19 02362 g003
Figure 4. Different distributions of the desired correlation output: (a) 2-D Gaussian distribution; (b) 2-D triangle distribution; (c) 2-D merged distribution.
Figure 4. Different distributions of the desired correlation output: (a) 2-D Gaussian distribution; (b) 2-D triangle distribution; (c) 2-D merged distribution.
Sensors 19 02362 g004
Figure 5. The results of the parameter setting experiments on OTB-2013 benchmark. The color bar denotes the one pass evaluation (OPE) accuracy percent (%) of location error threshold (LET) at 20 pixels. The horizontal and vertical coordinates represent merging factor α and learning rate η t 2 respectively. The points ranging from (0.1, 0.1), (0.1, 0.2), …, (1.0, 1.0) were obtained in experiments, others were obtained by the linear interpolation method. When the point ( α , η t 2 ) is set to (0.9, 0.5), the OPE accuracy is maximized.
Figure 5. The results of the parameter setting experiments on OTB-2013 benchmark. The color bar denotes the one pass evaluation (OPE) accuracy percent (%) of location error threshold (LET) at 20 pixels. The horizontal and vertical coordinates represent merging factor α and learning rate η t 2 respectively. The points ranging from (0.1, 0.1), (0.1, 0.2), …, (1.0, 1.0) were obtained in experiments, others were obtained by the linear interpolation method. When the point ( α , η t 2 ) is set to (0.9, 0.5), the OPE accuracy is maximized.
Sensors 19 02362 g005
Figure 6. The results of the ablation experiments on OTB-2013 benchmark for the baseline tracker, the proposed PCF tracker and the trackers of progressively integrating the presented strategies: (a) The precision plots (PP) of one pass evaluation (OPE) and the values in brackets denote the OPE accuracy of location error threshold (LET) at 20 pixels; (b) The success plots (SP) of the OPE and the numbers in brackets are the areas under the curve (AUC).
Figure 6. The results of the ablation experiments on OTB-2013 benchmark for the baseline tracker, the proposed PCF tracker and the trackers of progressively integrating the presented strategies: (a) The precision plots (PP) of one pass evaluation (OPE) and the values in brackets denote the OPE accuracy of location error threshold (LET) at 20 pixels; (b) The success plots (SP) of the OPE and the numbers in brackets are the areas under the curve (AUC).
Sensors 19 02362 g006
Figure 7. The quantitative results of the top ten trackers on OTB-2013 benchmark among three different contributions: DEF, OPR, and IV. Among the top ten trackers, the proposed PCF tracker obtains the best results on all three attributes: (a) The PP of the OPE on attribute DEF; (b) The SP of the OPE on attribute DEF; (c) The PP of the OPE on attribute OPR; (d) The SP of the OPE on attribute OPR; (e) The PP of OPE on attribute IV; (f) The SP of the OPE on attribute IV.
Figure 7. The quantitative results of the top ten trackers on OTB-2013 benchmark among three different contributions: DEF, OPR, and IV. Among the top ten trackers, the proposed PCF tracker obtains the best results on all three attributes: (a) The PP of the OPE on attribute DEF; (b) The SP of the OPE on attribute DEF; (c) The PP of the OPE on attribute OPR; (d) The SP of the OPE on attribute OPR; (e) The PP of OPE on attribute IV; (f) The SP of the OPE on attribute IV.
Sensors 19 02362 g007aSensors 19 02362 g007b
Figure 8. The comparing results for the top ten trackers with the whole 50 sequences on OTB-2013 benchmark: (a) The PP of the OPE on all sequences; (b) The SP of the OPE on all sequences.
Figure 8. The comparing results for the top ten trackers with the whole 50 sequences on OTB-2013 benchmark: (a) The PP of the OPE on all sequences; (b) The SP of the OPE on all sequences.
Sensors 19 02362 g008
Figure 9. The qualitative experiments from four representative sequences on the OTB-2013 benchmark. The results for the top ten trackers are expressed in different colors. On these challenging sequences, the proposed PCF tracker achieves remarkable results compared to the other top nine trackers in terms of accuracy and robustness.
Figure 9. The qualitative experiments from four representative sequences on the OTB-2013 benchmark. The results for the top ten trackers are expressed in different colors. On these challenging sequences, the proposed PCF tracker achieves remarkable results compared to the other top nine trackers in terms of accuracy and robustness.
Sensors 19 02362 g009
Figure 10. The quantitative results of the top ten trackers on OTB-2015 benchmark determined by three challenge contributions: DEF, OPR, and IV. Among the top ten trackers, the proposed PCF tracker achieves the top ranks in all three attributes: (a) The PP of the OPE on attribute DEF; (b) The SP of the OPE on attribute DEF; (c) The PP of the OPE on attribute OPR; (d) The SP of the OPE on attribute OPR; (e) The PP of the OPE on attribute IV; (f) The SP of the OPE on attribute IV.
Figure 10. The quantitative results of the top ten trackers on OTB-2015 benchmark determined by three challenge contributions: DEF, OPR, and IV. Among the top ten trackers, the proposed PCF tracker achieves the top ranks in all three attributes: (a) The PP of the OPE on attribute DEF; (b) The SP of the OPE on attribute DEF; (c) The PP of the OPE on attribute OPR; (d) The SP of the OPE on attribute OPR; (e) The PP of the OPE on attribute IV; (f) The SP of the OPE on attribute IV.
Sensors 19 02362 g010aSensors 19 02362 g010b
Figure 11. The comparing results for the top ten trackers with the entire 100 videos on the OTB-2015 benchmark: (a) The PP of the OPE on all videos; (b) The SP of the OPE on all videos.
Figure 11. The comparing results for the top ten trackers with the entire 100 videos on the OTB-2015 benchmark: (a) The PP of the OPE on all videos; (b) The SP of the OPE on all videos.
Sensors 19 02362 g011
Figure 12. The qualitative experiments from four representative videos on the OTB-2015 benchmark. The results for the top ten trackers are marked in different colors. In these challenging videos, the proposed PCF tracker performs better than the other top nine trackers.
Figure 12. The qualitative experiments from four representative videos on the OTB-2015 benchmark. The results for the top ten trackers are marked in different colors. In these challenging videos, the proposed PCF tracker performs better than the other top nine trackers.
Sensors 19 02362 g012
Figure 13. The failure cases from two representative sequences on the benchmarks. The tracking results for the top ten trackers are marked in different colors.
Figure 13. The failure cases from two representative sequences on the benchmarks. The tracking results for the top ten trackers are marked in different colors.
Sensors 19 02362 g013
Table 1. The parameter settings of the proposed PCF tracker.
Table 1. The parameter settings of the proposed PCF tracker.
Parameters η t 1 η t 2 α w k p 0 m
Values0.0090.50.91.030
Table 2. The PP of the OPE for the baseline tracker, the proposed PCF tracker and the trackers of progressively integrating the presented strategies on eleven different attributes (the ranks of results are expressed in %): scale variation (SV), illumination variation (IV), out-of-plane rotation (OPR), occlusion (OCC), background cluttered (BC), deformation (DEF), motion blur (MB), fast motion (FM), in-plane rotation (IPR), out-of-view (OV), and low resolution (LR). The last column is the average running speed (FPS) on all sequences of OTB-2013. The top three ranks are marked in red, cyan, and blue respectively.
Table 2. The PP of the OPE for the baseline tracker, the proposed PCF tracker and the trackers of progressively integrating the presented strategies on eleven different attributes (the ranks of results are expressed in %): scale variation (SV), illumination variation (IV), out-of-plane rotation (OPR), occlusion (OCC), background cluttered (BC), deformation (DEF), motion blur (MB), fast motion (FM), in-plane rotation (IPR), out-of-view (OV), and low resolution (LR). The last column is the average running speed (FPS) on all sequences of OTB-2013. The top three ranks are marked in red, cyan, and blue respectively.
TrackersSVIVOPROCCBCDEFMBFMIPROVLRSpeed
Baseline78.277.382.986.177.786.171.674.377.574.756.549
PCF_WF80.380.987.289.182.491.273.974.381.379.771.044
PCF_SR83.579.487.091.081.386.274.778.080.981.976.747
PCF82.482.188.489.584.591.376.076.782.781.173.342
Table 3. The SP of the OPE for the baseline tracker, the proposed PCF tracker and the trackers of progressively integrating the presented strategies on different attributes (%): SV, IV, OPR, OCC, BC, DEF, MB, FM, IPR, OV, LR. The last column is the AUC which represents the average overlap accuracy of the OPE. The top three ranks are marked in red, cyan, and blue separately.
Table 3. The SP of the OPE for the baseline tracker, the proposed PCF tracker and the trackers of progressively integrating the presented strategies on different attributes (%): SV, IV, OPR, OCC, BC, DEF, MB, FM, IPR, OV, LR. The last column is the AUC which represents the average overlap accuracy of the OPE. The top three ranks are marked in red, cyan, and blue separately.
TrackersSVIVOPROCCBCDEFMBFMIPROVLRAUC
Baseline60.359.561.664.458.565.157.757.858.160.135.864.3
PCF_WF61.261.864.565.961.068.958.157.560.463.444.566.2
PCF_SR62.860.464.267.460.365.658.359.059.664.748.166.0
PCF63.263.165.966.963.569.660.159.362.264.548.867.5
Table 4. The PP of the OPE on OTB-2013 benchmark for the top ten trackers on eleven different attributes (%): SV, IV, OPR, OCC, BC, DEF, MB, FM, IPR, OV, LR. The last column is the average precision (AP) of LET at 20 pixels. The top three ranks are marked in red, cyan, and blue, respectively.
Table 4. The PP of the OPE on OTB-2013 benchmark for the top ten trackers on eleven different attributes (%): SV, IV, OPR, OCC, BC, DEF, MB, FM, IPR, OV, LR. The last column is the average precision (AP) of LET at 20 pixels. The top three ranks are marked in red, cyan, and blue, respectively.
TrackersSVIVOPROCCBCDEFMBFMIPROVLRAP
PCF63.263.165.966.963.569.660.159.362.264.548.867.5
DeepSRDCF57.553.658.659.953.660.154.254.855.056.646.160.0
SiamFC_3s57.050.555.956.154.253.549.051.555.254.552.958.5
SRDCF54.151.555.157.453.058.651.651.451.751.141.057.4
Staple53.754.355.755.954.559.347.845.253.540.536.256.1
SAMF49.249.252.556.649.854.949.847.950.957.133.854.1
LCT50.449.353.852.053.958.642.340.851.448.434.854.1
FDSST52.151.752.150.757.351.448.847.552.347.928.854.0
DLSSVM42.045.949.951.249.856.652.448.948.949.335.951.5
DSST47.644.643.144.945.838.634.334.546.138.431.346.9
Table 5. The SP of the OPE on OTB-2013 benchmark for the top ten trackers on eleven different attributes (%): SV, IV, OPR, OCC, BC, DEF, MB, FM, IPR, OV, LR. The last column is the average overlap accuracy of the OPE. The top three ranks are marked in red, cyan, and blue respectively.
Table 5. The SP of the OPE on OTB-2013 benchmark for the top ten trackers on eleven different attributes (%): SV, IV, OPR, OCC, BC, DEF, MB, FM, IPR, OV, LR. The last column is the average overlap accuracy of the OPE. The top three ranks are marked in red, cyan, and blue respectively.
TrackersSVIVOPROCCBCDEFMBFMIPROVLRAUC
PCF82.482.188.489.584.591.376.076.782.781.173.389.2
DeepSRDCF80.776.584.685.976.784.174.975.680.078.274.985.3
SiamFC_3s76.768.076.375.673.170.964.768.274.768.378.079.3
SRDCF75.969.076.577.270.676.669.869.472.265.476.878.9
Staple75.373.577.076.373.877.962.760.473.465.472.276.5
SAMF73.469.075.980.765.674.064.863.072.966.870.877.4
LCT72.168.676.071.173.578.355.952.072.456.768.175.9
FDSST75.174.074.270.779.468.068.364.474.059.653.975.3
DLSSVM67.466.575.672.369.682.872.966.874.261.982.675.7
DSST68.260.060.060.159.849.143.242.165.148.669.364.3
Table 6. The PP of the OPE on the OTB-2015 benchmark for the top ten trackers on eleven different attributes (%): SV, IV, OPR, OCC, BC, DEF, MB, FM, IPR, OV, LR. The last column is the AP of LET at 20 pixels. The top three ranks are expressed in red, cyan, and blue respectively.
Table 6. The PP of the OPE on the OTB-2015 benchmark for the top ten trackers on eleven different attributes (%): SV, IV, OPR, OCC, BC, DEF, MB, FM, IPR, OV, LR. The last column is the AP of LET at 20 pixels. The top three ranks are expressed in red, cyan, and blue respectively.
TrackersSVIVOPROCCBCDEFMBFMIPROVLRAP
PCF83.982.585.184.482.883.082.382.480.779.189.286.3
DeepSRDCF80.880.082.983.478.778.779.479.381.981.188.084.2
SRDCFdecon75.676.272.871.574.272.474.272.171.359.678.277.3
SRDCF71.873.469.966.768.068.270.071.171.755.079.575.3
Staple73.276.472.669.971.471.064.967.775.759.277.375.0
SAMF68.470.171.069.863.667.459.865.071.957.376.973.5
LCT63.573.269.361.066.565.555.257.570.844.168.470.1
FDSST64.170.962.659.974.557.163.865.669.252.565.368.7
DSST61.862.157.153.762.249.751.451.964.150.168.863.3
KCF58.063.159.652.662.355.950.654.963.537.267.162.3
Table 7. SP of the OPE on OTB-2015 benchmark for the top ten trackers on eleven different attributes (%): SV, IV, OPR, OCC, BC, DEF, MB, FM, IPR, OV, LR. The last column is the average precision of LET at 20 pixels. The last column is the AUC of the OPE. The top three ranks are expressed in red, cyan, and blue respectively.
Table 7. SP of the OPE on OTB-2015 benchmark for the top ten trackers on eleven different attributes (%): SV, IV, OPR, OCC, BC, DEF, MB, FM, IPR, OV, LR. The last column is the average precision of LET at 20 pixels. The last column is the AUC of the OPE. The top three ranks are expressed in red, cyan, and blue respectively.
TrackersSVIVOPROCCBCDEFMBFMIPROVLRAUC
PCF61.663.662.661.663.062.563.261.459.356.256.064.7
DeepSRDCF56.557.056.857.555.753.658.957.854.654.150.558.7
SRDCFdecon54.355.752.152.755.052.257.254.549.948.046.755.9
SRDCF51.854.250.650.751.450.452.954.450.042.846.854.7
Staple51.355.451.651.352.352.350.450.751.445.041.153.7
SAMF46.049.249.348.947.546.947.348.948.944.839.351.0
FDSST46.449.645.744.354.242.847.649.649.641.539.150.0
LCT44.750.048.845.049.546.744.245.748.337.837.149.9
DSST43.345.141.240.446.137.241.041.644.739.033.845.6
KCF36.441.039.636.843.439.338.240.442.231.829.042.1

Share and Cite

MDPI and ACS Style

Yang, Y.; Zhang, Y.; Li, D.; Wang, Z. Parallel Correlation Filters for Real-Time Visual Tracking. Sensors 2019, 19, 2362. https://doi.org/10.3390/s19102362

AMA Style

Yang Y, Zhang Y, Li D, Wang Z. Parallel Correlation Filters for Real-Time Visual Tracking. Sensors. 2019; 19(10):2362. https://doi.org/10.3390/s19102362

Chicago/Turabian Style

Yang, Yijin, Yihong Zhang, Demin Li, and Zhijie Wang. 2019. "Parallel Correlation Filters for Real-Time Visual Tracking" Sensors 19, no. 10: 2362. https://doi.org/10.3390/s19102362

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop