Next Article in Journal
The Gas Fire Temperature Measurement for Detection of an Object’s Presence on Top of the Burner
Previous Article in Journal
Detection of Atrial Fibrillation Using 1D Convolutional Neural Network
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

HKSiamFC: Visual-Tracking Framework Using Prior Information Provided by Staple and Kalman Filter

College of Biomedical Engineering and Instrument Science, Zhejiang University, Hangzhou 310027, China
*
Author to whom correspondence should be addressed.
Sensors 2020, 20(7), 2137; https://doi.org/10.3390/s20072137
Submission received: 22 February 2020 / Revised: 27 March 2020 / Accepted: 8 April 2020 / Published: 10 April 2020
(This article belongs to the Section Physical Sensors)

Abstract

:
In the field of visual tracking, trackers based on a convolutional neural network (CNN) have had significant achievements. The fully-convolutional Siamese (SiamFC) tracker is a typical representation of these CNN trackers and has attracted much attention. It models visual tracking as a similarity-learning problem. However, experiments showed that SiamFC was not so robust in some complex environments. This may be because the tracker lacked enough prior information about the target. Inspired by the key idea of a Staple tracker and Kalman filter, we constructed two more models to help compensate for SiamFC’s disadvantages. One model contained the target’s prior color information, and the other the target’s prior trajectory information. With these two models, we design a novel and robust tracking framework on the basis of SiamFC. We call it Histogram–Kalman SiamFC (HKSiamFC). We also evaluated HKSiamFC tracker’s performance on dataset of the online object tracking benchmark (OTB) and Temple Color (TC128), and it showed quite competitive performance when compared with the baseline tracker and several other state-of-the-art trackers.

1. Introduction

Visual tracking plays a fundamental role in the field of computer vision. It is widely used, including in auto-monitoring [1], autonomous driving [2], surveillance [3], robotics [4], and motion-based recognition [5]. The main task of visual tracking is distinguishing a target from complicated environments in video, and to provide an accurate result of the target size and location co-ordinates [6], usually with a bounding box. The only reference that the tracker can use is the labeled ground truth in the first frame. This seemingly easy task has met many difficulties. Lacking prior information about the target’s appearance, environment illumination changes [7,8,9], scene clutter [10], and many other factors can lead to the tracker drifting, and make tracking fail. To solve these challenging problems, a large variety of novel trackers have been proposed over the years. They are constructed on different theoretical bases and aimed at solving different challenges. However, all those trackers share the same key step: extracting powerful features from the target that are distinct enough to express target properties. Then, those features can be used as a reference to locate the right target and leave out extraneous distractors.
Recently, deep learning (DL) has flourished and benefited many computer-vision-related tasks [11,12,13,14,15]. A convolutional neural network (CNN) [16] is a typical deep-learning architecture that is often used in image processing. Unlike traditional features, a CNN can use many training data and learn to capture target features from different levels. The number of levels is decided by the CNN’s number of layers, and each layer acts as a specific feature-pattern extractor. So, features extracted by CNN are very powerful and widely used by many trackers.
Typically, Bertinetto designed a fully-convolutional Siamese (SiamFC) tracker [17] using Alexnet [18]. It achieved state-of-the-art performance on the visual-tracking benchmark, showing its significant potential in visual tracking. SiamFC models visual tracking as a similarity learning problem, and the CNN in SiamFC is trained end-to-end especially for visual tracking. This determines that the features it extracts are more powerful than other CNNs which are trained in other tasks. For example, the deep features used in many correlation filter trackers are extracted by CNNs which are trained in image detection task.
However, there are two main disadvantages in the SiamFC tracking framework. First, it only uses the features in the last layer to generate the final response scores. Those high-level features are robust to the noise but lack detail information of the target, so they are not discriminative enough when the distractor belongs to the same category with the target. Second, the way to train the SiamFC determines that all the patches in the search region have the same contribution to the final response score map. So, when a distractor appears in the search region, no matter where its location is, it may generate a high response score and cause the tracking to fail.
Our research focuses on addressing the problems listed above. We find that these two problems essentially show the same key point that SiamFC lacks enough prior information about the target. So our motivation is to provide prior information about the target to the original SiamFC tracker, containing appearance information and motion information. As for the first problem, the high-level features’ shortcoming can be complemented by adding more prior detail information about the target’s appearance, and we think color information is a good choice because color distribution of the target copes well with appearance variation. We analysis the histogram score model used in Staple and find it can be improved to achieve our goal here. The other problem can be improved by adding prior trajectory information of the target. Then, we can use the trajectory information to generate a Gaussian likelihood map. Then, the response score of the distractor far from the target can be suppressed by this map. The Kalman filter is a good choice for us to get the prior trajectory information we need. In this paper, we design a novel tracking framework based on the motivation above. The framework contains three main components, namely SiamFC, histogram score model, and Kalman filter model. This tracker is called Histogram–Kalman SiamFC (HKSiamFC). The main contributions of our contributions are summarized as follows:
(1)
We improve the Staple’s histogram score model and introduce it into our HKSiamFC tracker. Unlike the Staple, we don’t use the histogram score model to generate an integral image, but optimize it with Gaussian filter and use it to provide appearance information of the target.
(2)
We build a Gaussian likelihood map on the basis of a Kalman filter to describe the target’s trajectory information, and then use this Gaussian likelihood map and the histogram score map to process the SiamFC’s response score map by element-wise multiplication.
(3)
We evaluated HKSiamFC tracker on the dataset of online tracking benchmark (OTB) and Temple Color (TC128), and experiments showed that it achieved competitive performance when compared with baseline trackers and other state-of-the-art trackers.
The rest of the paper is organized as follows. In Section 2, we introduce some related works in visual tracking. Then, we present our tracking framework in Section 3. In Section 4, we evaluate the performance of our tracker on the mainstream dataset in visual tracking and compare it with some state-of-the-art trackers. Section 5 presents a summary of our work.

2. Related Work

2.1. Correlation-Filter-Based Trackers

Most correlation filter-based trackers have three characteristics in common: learning a correlation filter (CF), cyclic sampling, and fast Fourier transform. The first CF-based tracker came from Bolme et al. [19]. They designed the tracker with a minimal output sum of a squared error filter (MOSSE) using raw image pixels as inputs to train the correlation filters without any feature extraction. After Bolme, Henriques et al. [20] designed the circulant structure kernel (CSK) tracker that modeled the training of a correlation filter as a ridge regression problem, and introduced a circulant structure and kernel tricks to optimize the tracking framework. However, the CSK tracker only used greyscale features to train the filter, and this largely limited its performance. To overcome this disadvantage, Danelljan et al. [21] integrated the image‘s color properties into the CSK tracker, and experiments showed that this method significantly improved its performance. Then, with the kernelized correlation filter (KCF)/discriminative correlation filter (DCF) tracker [22] designed by Henriques et al., the process of feature extraction was further optimized. They extended the target’s feature representation into multichannel histograms of oriented gradients (HOG) and color name (CN) features instead of single-channel features, and introduced a Gaussian kernel to accelerate tracker efficiency. The long-term correlation tracker (LCT) [23] is a typical long-term tracker using CF that concentrates on solving the problem of target deformation during the tracking process, and decomposes the tracking task into two parts, translation estimation and scale estimation. The discriminative scale space tracker (DSST) [24] is another CF-based tracker that focuses on tackling the target’s fast-scale estimation. It trains a translation filter and a scale filter using handcrafted features, respectively. The background-aware correlation-filter (BACF) tracker [10] indicated that a robust tracker not only needs to boost its response to the target, but also needs to learn how to suppress its response to the background. So, this tracker utilizes background and target-variation information to build a discriminative classifier. The Staple [25] tracker uses the target‘s color-statistics information to build a histogram-score model and combine it with the model that is learned on the spatial layout of target, showing robust and balanced performance in many complex environments. The adaptive spatially-regularized correlation filter (ASRCF) [26] tracker designed a novel model which could optimize the spatial regularization weight and the filter coefficients at the same time.

2.2. Deep-Learning-Based Trackers

The main characteristic of deep-learning-based trackers is that those trackers introduce a deep-learning (DL) framework into visual tracking. This characteristic can be viewed from two aspects. A deep-learning framework (especially a CNN) can be used as a powerful feature extractor. We can train a deep neural network with a large amount of data so that the network learns to extract features from different levels. Sometimes, a CNN can be integrated into a CF tracking framework as a feature-extractor component. For example, the deep spatially regularized discriminative correlation-filter (SRDCF) tracker [27] trains an CNN and employs extracted features from its shallow layers in a spatially regularized DCF tracker. The spatial temporal regularized correlation filter (STRCF) tracker [28] introduced temporal regularization into SRDCF with a single sample. The hedged deep tracker (HDT) [29] and hierarchical convolutional features (HCF) tracker [30] are also typical trackers using a CNN as feature extractor.
The other aspect of using a deep-learning framework in visual tracking is to bring its end-to-end characteristics into full play. As discussed in Section 1, SiamFC [17] is a typical tracker of this category. This framework realized end-to-end workflow specifically for tracking problems and achieved quite competitive performance. On the basis of the SiamFC tracker, CFNet [31] integrated a correlation filter into SiamFC and interpreted the CF as a differentiable CNN layer. In this way, the CF could be trained end to end. The dynamic Siamese (DSiam) tracker [32] constructed a component which could simultaneously represent the target’s appearance and suppress background noise. The SiamRPN [33] tracker added a regression branch to the original SiamFC, and introduced a region proposal network (RPN) [34] into visual tracking. So, this tracker could largely improve its performance in target-scale estimation. DiaSiamRPN [35] tracker introduced a novel distractor-aware component on the basis of SiamRPN and improved its performance. SiamFC-tri [36] tracker designed a new triplet loss function in SiamFC to extract more powerful features. Deeper and wider SiamFC (DWSiamFC) tracker [37] analysed key factors of the CNN that influenced the tracker’s performance and showed how to make the network deeper and wider. The cascaded RPN (C-RPN) tracker [38] introduced hard negative sampling and designed the cascaded RPN to address the problem of training data imbalance. However, few trackers tried using prior information to guide the tracking process.

3. Proposed Algorithm

The tracking framework we propose in this paper consists of three main components, namely, SiamFC, a histogram-score model from Staple, and a Gaussian likelihood model based on a Kalman filter. Our main purpose was to use the histogram-score and Gaussian likelihood models to suppress the background’s response score. When the SiamFC component outputs a response score map, our tracker will use the histogram score map and the Gaussian likelihood map to process the response score map by element-wise multiplication. The overview of our algorithm’s workflow is shown in Figure 1, and it is described in detail in the following sections.

3.1. SiamFC Tracker

The SiamFC tracker models visual tracking as a similarity-learning problem, and trains the twin neural network to learn a general similarity-matching function. Its main architecture is made of two CNN branches that share the same parameters and are used to process the template patch and the searching-area patch, respectively. The template patch is initially set as the ground truth from the first frame. The searching-area patch is cropped from the current frame that is to be searched for the target. The template and searching-area patches are denoted as z and x, respectively. Two patches x and z were input into Alexnet and generated the feature maps denoted as φ ( x ) and φ ( z ) . As the size of φ ( z ) was smaller, it was used as a sliding window on φ ( x ) to evaluate the similarity scores between z and each subregion of x. The similarity score was computed by the formulation below:
S i m i S c o r e ( x ( i ) ,   z ) = ( m ,   n ) S φ ( z ) ( m ,   n ) × φ ( x ( i ) ) ( m ,   n ) ,  
where x(i) denotes the i th subregion of x that had the same size as z, φ ( z ) ( m ,   n ) denotes the value vector of the pixel located at co-ordinates (m, n) on the feature maps of z, and S represents the whole area of each feature map. So, φ ( z ) slid onto φ ( x ) from the top left to the bottom right and produced a score map, and each pixel on it showed the similarity between the subregion of x and template z. Then the subregion that had the maximal value was selected as the target.
AlexNet in SiamFC was trained by an offline pretraining method with a logistic loss function. Training data were positive and negative pairs of images. Those images were extracted from two frames of the same video that contained a target. Each pixel value of the score map used for training was labelled as +1 or −1 on the basis of its distance to the map center. If the distance were smaller than a threshold, the pixel was labelled as +1; otherwise, it was labeled as −1. The logistic loss function adopted here was defined as
l o s s ( y ,   r ) = l o g ( 1 + e x p ( y r ) ) ,  
where y is the ground-truth binary score map, and r is the real-valued score map. So, the final loss of score map r is defined as the mean of each pixel’s loss:
L ( y ,   r ) = 1 | D | u D l ( y [ u ] ,   r [ u ] ) ,  
where D denotes the whole score map, and u D represents each pixel on it. AlexNet parameters were achieved by using the stochastic-gradient-descent (SGD) method. For more details about this issue, please refer to [17].

3.2. Histogram Score Model from Tracker Staple

The key idea behind the Staple tracker is the intuition that an object’s color distribution is always stable and discriminative enough, no matter how much the object’s appearance changes. So, a model based on color statistics can provide powerful information for visual-tracking problems. The information that this model offers can be seen as prior information, and this information is what SiamFC lacks. So, we integrated this histogram-score model into HKSiamFC.
To calculate the histogram score of an image, an N-Channel weight map is introduced; here, N is the number of input feature maps. That is to say, for each pixel on each feature map, there is a weight value provided for it to compute the histogram score. Here, this weight map is denoted as β , and this issue is formulated as
f h i s t ( x ; β ) = h ( τ x ; β ) ,  
where x is the original image, and τ x is an N-channel feature map extracted from x, having the same channel number as β . Then, τ x and β were combined to calculate the histogram score. As the histogram score should be invariant to spatial permutations of τ x , i.e., h ( τ ) = h ( τ ) should be satisfied for any permutation matrix Π , so h can be formulated as
h ( τ ; β ) = β T ( 1 | ϱ | u ϱ τ [ u ] ) ,  
where   ϱ represents the whole area of each feature map, and τ [ u ] represents each pixel on ϱ . The way to compute β is to model this problem as a linear-regression problem. For any pixel that belongs to the target, the ground-truth label of β T τ [ u ] should be 1, and for any pixel that belongs to the background, the ground-truth label of β T τ [ u ] should be 0. Then, the loss function can be formulated as
1 | B | u B ( β T τ [ u ] ) 2 + 1 | O | u O ( β T τ [ u ] 1 ) 2 ,  
where O represents the area of the target region, and B represents the area of the background region. As the number of feature maps is N, and every feature map is independent of the others, this formulation can be decomposed into independent terms for each feature dimension, so that
j = 1 N [ P j ( O ) | O | ( β j 1 ) 2 + P j ( B ) | B | ( β j ) 2 ] ,  
where N j ( O ) denotes the number of pixels in region O for which feature dimension j is non-zero k [ u ] = j .Then, the solution of this regression problem is
β t j = ρ j ( O ) ρ j ( O ) + ρ j ( B ) + λ ,   j = 1 ,   2 ,   3     N ,  
where ρ j ( O ) is the proportion of pixels in a region for which feature j is non-zero. For more details about this issue, please refer to [25].

3.3. Optimize Histogram Score with Gaussian Filter

In Reference [25], weight map β was used to generate the integral map of an image. In this paper, we used β to optimize the final score map generated from SiamFC. We were inspired to do so when we saw the histogram-score map of the Human8 sequence from OTB during the experiments. As shown in Figure 2, the histogram score of the target region was larger than that of the background region (pixels having a larger histogram are marked with red). The SiamFC tracker failed badly in this sequence, even from the first few frames. The histogram-score map could have compensated for SiamFC’s failure, so we combined SiamFC’s response-score map with Staple’s histogram-score map using element-wise multiplication; this process could suppress the background-response score so that the response of the right target can be more prominent.
As discussed in Section 3.1, the size of search-region patch x in SiamFC was 255 × 255 pixels, and the size of the final response-score map was 17 × 17 pixels. To locate the target co-ordinates, interpolation was used to resize the final response-score map to 255 × 255 pixels, the same size as the search region. In our HKSiamFC tracker, we generated another histogram-score map with a size of 255 × 255 pixels following the method introduced in [25]. However, we found that this map could not be directly used for element-wise multiplication because it contained many isolated pixels that belong to the noise but still had a higher histogram score. In Staple, the influence of these noise pixels is eliminated by generating an integral image. However, as we did not use an integral image in our method, we adopted the Gaussian filter to address this problem.
A Gaussian filter is essentially a low-pass filter that calculates image convolution and normal distribution. It can remove fine image details, and produce a very pure smoothing effect. According to our experiments, in our HKSiamFC tracker, a Gaussian filter with 9 × 9 kernel size could eliminate the bad influence from the isolated noise pixels very well, as shown in Figure 3.
As this histogram-score model was constructed on the basis of the image’s RGB color information, and experiments showed that greyscale features weaken its performance, we added a component to judge whether the image sequence was greyscale or not. The histogram model was only used when the image is in RGB color. This component made our tracker more robust.

3.4. Gaussian Likelihood Map Based on Kalman Filter

Both SiamFC and Staple’s histogram model have the same disadvantage in visual tracking: they mainly concentrate on the appearance feature of the target while ignoring its trajectory trait between frames, which is also a very important indication to locate the target. When a distractor that has a very similar appearance to the target occurs, the tracker could drift to the distractor, and tracking fails. Although SiamFC has limited the search region to a small patch of the whole image, the influence of the distractor still exists. To solve this problem, we introduced a Gaussian likelihood map based on a Kalman filter.
A Kalman filter is essentially an optimal linear-estimation algorithm used for dynamic systems. It is based on the hidden Markov model and provides a recursive solution for time-series analysis. Usually, this dynamic system can be represented by a Markov chain that contains Gaussian noise. The purpose of the Kalman filter is to predict unknown states { X t ,   t = 1 ,   2 ,   3 n } from a series of measurements { Z t ,   t = 1 ,   2 ,   3 n } using the following formulations:
X t = A t ,   t 1 X t 1 + W t ,  
Z t = C t X t + V t ,  
where A is the state-transition model, and A t ,   t 1 denotes the transition coefficients from time t 1 to time t; C is the observation model, and C t denotes the observation at time t; and W t and V t are defined as the white noise at time t that follows Gaussian distribution. Equation (9) is called the state equation, and Equation (10) is called the measurement equation. In our proposed HKSiamFC tracker, a Kalman filter is used to estimate the target’s location co-ordinates in every frame.
The workflow of the Kalman filter has two main steps, state prediction and measurement update. When a Kalman filter is used to estimate the target’s location, its state matrix consists of the target position co-ordinates and velocity. In the prediction step, the state equation is used to predict prior state x t 1 | t 1 and the error’s prior covariance P t of the t-th frame by
x t | t 1 = F t 1 x t 1 ,  
P t | t 1 = F t 1 P t 1 | t 1 F t 1 T ,  
where P t 1 | t 1 is the covariance of previous state x t 1 | t 1 . Then, in the update step, the measurement is used to refine the state prediction. This step can be described by the formulations:
K t = P t | t 1 H T ( H P t H T + R t ) 1 ,  
P t = P t | t 1 ( I K t H ) ,  
x t | t = x t | t 1 + K t ( z t z t | t 1 ) ,  
where K t is the Kalman gain at time t, P t is the updated covariance, and I in Equation (14) denotes the identity matrix. x t | t is the posterior estimation of the t-th frame’s state. Finally, we can obtain the optimal estimation of the target’s location in the current frame from x t | t .
As discussed in Section 4, neither SiamFC nor the histogram-score map has the ability to use trajectory information to distinguish targets. This means that distractors in the background can still have high response scores if they have a similar appearance to the target. After obtaining a prior estimation of the target location from the Kalman filter, this situation can be improved. We could use the estimated location as the peak to build a two-dimensional Gaussian likelihood map, and the standard deviation of this distribution is computed as
σ = w × h λ ,   λ { 0.16 ,   1.6 ,   16 } ,  
where w and h are the width and length of the target that are provided by the previous frame’s tracking result, respectively, while λ is a parameter. Figure 4 shows the workflow of Kalman model. This map shows the probability of each pixel on the image belonging to the target using trajectory information provided by the Kalman filter. So, utilizing it for element-wise multiplication with SiamFC’s score map can suppress background noise from the perspective of motion information.

3.5. Template Update

In the course of visual tracking, the target’s appearance and its surroundings are always changing as tracking goes on, so only using the annotated ground truth in the initial frame as the template is not enough—the template needs to be properly updated over time. A template update can provide more online information about the target, and make the tracker more robust. In the HKSiamFC tracker, the update strategy is formulated as
{ S ( t ) = ( 1 R s ) × S ( t 1 ) + R s × S ( t ) O ( t ) = ( 1 R o ) × O ( t 1 ) + R o × O ( t ) B ( t ) = ( 1 R b ) × B ( t 1 ) + R b × B ( t ) ,  
where S ( t ) ,   O ( t ) ,   B ( t ) represent the SiamFC feature template, target histogram feature template, and background histogram feature template updated in the t-th frame; R is the update coefficient. In our HKSiamFC tracker, all the update coefficients are set as 0.04.

4. Experiments

In this section, we evaluate the HKSiamFC tracker’s performance using the dataset of Online Object Tracking Benchmark (OTB) and Temple Color (TC128). Comprehensive experiments were performed to show the effectiveness of our tracking framework. The HKSiamFC tracker was implemented using Google’s TensorFlow library with Python. The platform we used to perform all the experiments was a Dell Alienware DESKTOP-N7K2SPB with a 3. 70 Ghz Intel Core i7-8700K CPU and an NVIDIA GeForce GTX 1080Ti GPU. The operating system was 64 bit Windows 10 Professional. Our HKSiamFC tracker can realize real-time tracking with a speed of 18 FPS.

4.1. Benchmark and Evaluation Metric

OTB [39] is a famous benchmark database that was designed especially for visual tracking. It contains 100 fully labeled image sequences. The number of frames that each sequence contains range from the hundreds to the thousands. Those sequences are marked with 11 attributes according to the tracking environment, and those attributes contain illumination variation, low resolution, fast motion, background clutter, in-plane rotation, deformation, occlusion, out-of-plane rotation, motion blur, out of view, and scale variation. More concretely, the 100 sequences are divided into three subsets: OTB100 (OTB2015), OTB50, and OTB2013. OTB100 contains all those 100 sequences, while OTB50 and OTB2013 both consist of 50 video sequences that are selected from OTB100. The OTB adopts the one-pass-evaluation (OPE) protocol to evaluate tracker performance, which means that it uses the tracker to process each image sequence from beginning to end only once, and then records the tracking results. TC128 [40] is another famous dataset which is designed for visual tracking evaluation. It consists of 128 sequences of color images and those images contains all kinds of complicated tracking environments. The evaluation protocol of TC128 is the same with OTB.
In our experiments, a uniform evaluation metric was adopted, put forward by the OTB and widely acknowledged in the community of visual tracking, namely, precision plot and success plot. These two plots are based on center location error (CLE) and intersection over union (IOU), respectively. The CLE compared the Euclidean distance between the center locations of the tracked result and the corresponding labeled ground-truth of each frame. Then, a threshold was set to determine whether tracking was successful. If Euclidean distance was smaller than the threshold, tracking was regarded as correct. Usually, different thresholds are set to more roundly evaluate trackers. The IOU is defined as
IOU = area ( R T R G ) a r e a ( R T R G )   ,  
where ∩ and ∪ denote the intersection area and union area between tracked results R T and labeled ground truth R G , respectively. Thus, the precision plot shows the percentage of successfully tracked frames on the basis of CLE, and concentrates more on location co-ordinate correctness, while the success plot shows the percentage of successfully tracked frames on the basis of IOU, and concentrates more on bounding-box size. The value representing the tracker’s performance in success plot is the Area Under Curve (AUC) value.

4.2. Ablation Experiment

In this section, an ablation experiment was conducted on OTB dataset to test the correctness and effectiveness of each tracking component proposed in this paper. We adopted SiamFC and Staple as our baseline trackers because our tracker was designed on that basis. These two trackers are representative state-of-the-art trackers that are cited often by visual-tracking researchers. The codes of SiamFC and Staple were downloaded from the authors’ website without any changes. The parameters of the trained CNN used in the SiamFC tracker came from “2016-08-17.net.mat” on http://www.robots.ox.ac.uk/~luca/siamese-fc.html. What’s more, to test the function of the histogram model, the Kalman model and the template update strategy of our tracking framework, we also set another three baseline trackers: (1) HSiamFC (without K): this tracker is produced by removing the Kalman model from the HKSiamFC. (2) KSiamFC (without H): this tracker is produced by removing the histogram model from the HKSiamFC. (3) HKSiamFC (without template update): this tracker is produced by removing the template update process from HKSiamFC.

4.2.1. Overall Performance

Figure 5 shows the overall performance of the four trackers on the three OTB datasets in terms of precision based on CLE and success based on IOU. The HKSiamFC tracker obviously outperformed all five other trackers, no matter which dataset was used. HKSiamFC obtained average precision scores of 86.4%, 85.9% and 88.0% on OTB100, OTB50, and OTB2013, respectively. When compared with the SiamFC tracker, our HKSiamFC tracker achieved obvious gains of 7.4%, 13.2%, and 4.6% in precision. When compared with Staple, the HKSiamFC tracker achieved precision gains of 8.9%, 18.4%, and 8.7%. With regard to the success plot, HKSiamFC obtained average scores of 62.2%, 59.0%, and 64.4% on OTB100, OTB50, and OTB2013, respectively. When compared with the SiamFC tracker, our HKSiamFC tracker achieved obvious success gains of 3.5%, 6.1%, and 2.0%. When compared with Staple, the HKSiamFC tracker achieved success gains of 4.5%, 8.4%, and 4.4%. In the precision plot of OTB2013, the performance of HSiamFC (without K; 0.813) was inferior to that of SiamFC’ (0.834). The same occurred in all three success plots. As for the KSiamFC(without H) tracker, we can see that its performances are inferior to SiamFC in all the plots. So only using Kalman model can’t give a stable boost to SiamFC. What’s more, in all the six plots, the HKSiamFC(no-template-update) tracker’s performance was the worst. This situation shows the effectiveness and necessity of template update. As the appearance of the target is keeping changing, the template must be updated properly as well to keep the tracker robust. In conclusion, according to the six trackers’ overall performance on the OTB dataset, HKSiamFC showed superior performance in precision and success rate compared with its baseline trackers. Improvement was achieved because the histogram-score model could provide powerful prior information about target appearance, and the Kalman model could provide powerful prior information about target trajectory. This prior information can effectively guide the original SiamFC and compensate for its disadvantage of treating every search region equally.

4.2.2. Scenario-Based Performance

As was discussed above, image sequences in OTB are marked with 11 different labels that describe the tracking scenarios that each sequence contains. Each label represents one crucial factor that may influence tracker performance, i.e., scale variation (SV), low resolution (LR), illumination variation (IV), motion blur (MB), out-of-plane rotation (OPR), out of view (OV), background cluttered (BC), deformation (DEF), fast motion (FM), in-plane rotation (IPR), and occlusion (OCC). Evaluating tracker performance under each scenario could therefore further analyze the tracker’s characteristics and find its advantages and disadvantages. In this way, we can better optimize the tracking algorithm.
Figure 6 and Figure 7 show the precision plots and the success plots of the six trackers’ tracking performances in the 11 tracking scenarios on OTB100. Precision values and AUC values of the six trackers in each scenario are summarized in detail in Table 1 (performance ranking first is marked with red color). In the low-resolution sequences, HKSiamFC’s AUC value (0.511) was inferior to that of SiamFC (0.528) by gaps of 0.017. In all the other tracking scenarios, HKSiamFC’s performance was better than that of all five other baseline trackers, obviously achieving varying degrees of improvements. This proves that our tracking strategy brings comprehensive optimization to the baseline trackers.

4.3. Comparison with State-of-the-Art Trackers on OTB100

To further evaluate the performance of our HKSiamFC tracker, we selected 14 state-of-the-art trackers designed on the basis of different tracking strategies as our comparison objects. Those trackers were all representative, and have been cited by many visual-tracking researchers. Those trackers are: kernelized-correlation-filter (KCF) tracker [22], multi-domain convolutional neural networks (MDNet) tracker [41], fully-convolutional Siamese (SiamFC) tracker [17], siamese network using triple loss (SiamFC_tri) tracker [36], efficient convolution operators (ECO-HC) tracker [42], context-aware deep feature compression with multiple Auto-encoders (TRACA) tracker [43], correlation filter network using triple loss (CFNet2_tri) tracker [36], cropping-inside residual fully-convolutional network (CIResNet-22FC) tracker [37], Staple tracker [25], spatially regularized discriminative correlation-filter (SRDCF) tracker [27], background-aware correlation-filter (BACF) tracker [10], discriminative correlation-filter network (DCFNet) tracker [44], large-margin correlation-filter (LMCF) tracker [45], hedged deep tracker (HDT) [29]. The selection of these trackers provided a horizontal comparison to comprehensively evaluate our HKSiamFC tracker. To be concise, we only listed the results of the precision and success plots on OTB100 because it contains all the sequences that OTB50 and OTB2013 have, so these results were persuasive enough.

4.3.1. Overall Performance

We compare our HKSiamFC tracker’s overall tracking performance with those of the 14 state-of-the-art trackers on OTB100. Figure 8 shows the results of the precision plot and success plot. HKSiamFC achieved the second-best performance in precision plot, only inferior to MDNet and outperforming all the other 13 state-of-the-art trackers. According to [41], the tracking speed of MDNet is 1 fps on GPU, so it is not actually a real-time tracker. HKSiamFC’s tracking speed (18 FPS) was far higher than that of MDNet. In the success plot, HKSiamFC achieved the fourth-best performance, outperforming 11 state-of-the-art trackers. So, our HKSiamFC’s overall performance on OTB100 is quite competitive within all 15 compared trackers.

4.3.2. Scenario-Based Performance on OTB100

We also compared HKSiamFC’s performance with that of 14 state-of-the-art trackers in 11 different scenarios on OTB100. Figure 9 shows the detailed results of all trackers’ performance in the precision plot. Table 2 shows HKSiamFC’s rankings of all scenarios in precision. We can see that HKSiamFC ranks second in six scenarios (BC, DEF, IPR, OCC, OPR, SV), more than half of the total 11 scenarios. This proves that our HKSiamFC tracker can handle many complicated tracking environments and shows competitive performance when compared with the state-of-the-art trackers.

4.4. Comparison with State-of-the-Art Trackers on TC128

We also conducted comparison experiments on the TC128 dataset. This dataset contains more image sequences than OTB100, so those experiments can show the effectiveness of our tracking strategy more comprehensively. We compared our HKSiamFC tracker’s performance with the baseline SiamFC tracker and many other classic trackers whose tracking results could be downloaded from the homepage. Figure 10 shows the precision plot and success plot of their tracking results. We can see that Our HKSiamFC tracker’s performance (0.7784, 0.5501) is obviously the best among all the trackers in both plots.
What’s more, there are also some other state-of-the-art trackers with published average precision and success values on TC128 dataset, namely: multi-task correlation particle filter (MCPF) [46] tracker, spatially regularized discriminative correlation-filter (SRDCF) tracker [31], background-aware correlation-filter (BACF) tracker [10], hedged deep tracker (HDT) [32], convolutional networks without training (CNT) tracker [47], adaptive decontamination of spatially regularized discriminative correlation filter (SRDCFdecon) tracker [48], convolutional features for correlation filter (DeepSRDCF) tracker [49]. We also compared our HKSiamFC tracker’s performance with those trackers and details were showed in Table 3. The first and second best scores are marked with red and blue colors, respectively. We can see that HKSiamFC’s precision (77.8) ranks first in all nine trackers. HKSiamFC’s AUC value (55.1) ranks second, only inferior to MCPF by a gap of 0.1. All the experiments proves that Our HKSiamFC’s performance is still competitive on TC128 dataset.

4.5. Qualitative Experiments

In this section, we present the qualitative tracking results of 11 trackers listed above on several representative sequences in the OTB dataset. These 11 trackers were HKSiamFC and 10 other trackers selected from the comparison objects in Section 4.3, details of which are shown in Figure 11. The six rows of sequences (from top to bottom) are biker, diving, human3, jump, skating2-1, and soccer, respectively.
In the Biker sequence, the most challenging moment was when the biker abruptly jumped from left to right; trackers are very easily led to drift by this fast motion before the man jumps. As is shown in the second image, some trackers already drifted into the background, and after the biker had jumped to the right, as is shown in the last three images, all other trackers drifted, while HKSiamFC is the only one that still precisely captured the target. The diving sequence is a typical sequence that contains background clutter (BC) and fast motion (FM); the audience in the background can lead the trackers to easily drift. Before the diver jumped into the river, some trackers had already failed, as is shown in the second image. While the diver was entering the river, as is shown in the fourth image, HKSiamFC was the only tracker that captured the target. In the human3 sequence, there were some other pedestrians besides the target. Those pedestrians, and even the lamppost, are powerful distractors that could mislead trackers, as is shown in the fifth image—only three trackers still captured the target, with HKSiamFC being one of them. The jump sequence is another that contains background clutters and fast motion. As is shown in the last three images of this sequence, HKSiamFC was the only tracker that captured the target while the jumper was landing. The skating2-1 sequence contains two people who constantly exchange positions and block each other, and this is the biggest challenge in this sequence. Trackers can very easily drift to the man skater, even though the woman skater is our right target. Figure 6 shows that our HKSiamFC tracker could constantly precisely capture the right target. The soccer sequence is a typical sequence of an occlusion scenario; the pieces of red paper are everywhere in the third and fourth images, and this badly influences tracker performance. We can see from the last three pictures that HKSiamFC could always capture the target, while many other trackers had already failed.

5. Conclusions

In this paper, we introduced prior information of the target into the SiamFC tracker and proposed the HKSiamFC tracker. Two components were used in HKSiamFC to generate prior information: the histogram-score component from Staple to generate prior information of the target’s appearance, and the Kalman filter component to generate prior information of the target’s motion. Comparative experiments with other state-of-the-art trackers were conducted to verify the effectiveness of HKSiamFC, and results showed that HKSiamFC’s performance was quite robust in many challenging tracking environments. However, we also observed that the bounding box of the tracking result given by our tracker sometimes contained too much background information, and this could influence its precision in the success plot. So, in the future, our work will mainly focus on optimizing the tracker with regard to this issue.

Author Contributions

C.L. constructed the tracking framework and performed the experiments and wrote the original manuscript. Q.X. and Z.M. analysed and interpreted the experiment results, and provided suggestions about the revision of this manuscript. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Ward, J.; Antonucci, G.; Barter, E.; Brooks-Tyreman, F.; Connaughton, C.; Coughlan, M.; Kuhne, R.; Kaiser, M.; Wang, V. Evaluation of the Accuracy of a Computer-Vision Based Crowd Monitoring System; CrowdVision: London, UK, 2018. [Google Scholar]
  2. Simon, M.; Amende, K.; Kraus, A.; Honer, J.; Samann, T.; Kaulbersch, H.; Milz, S.; Michael Gross, H. Complexer-YOLO: Real-Time 3D Object Detection and Tracking on Semantic Point Clouds. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Long Beach, CA, USA, 15–21 June 2019. [Google Scholar]
  3. Bouchrika, I.; Carter, J.N.; Nixon, M.S. Towards automated visual surveillance using gait for identity recognition and tracking across multiple non-intersecting cameras. Multimed. Tools Appl. 2016, 75, 1201–1221. [Google Scholar] [CrossRef]
  4. Tokekar, P.; Isler, V.; Franchi, A. Multi-target visual tracking with aerial robots. In Proceedings of the 2014 IEEE/RSJ International Conference on Intelligent Robots and Systems, Chicago, IL, USA, 14–18 September 2014; pp. 3067–3072. [Google Scholar]
  5. Lien, J.; Olson, E.M.; Amihood, P.M.; Poupyrev, I. RF-Based Micro-Motion Tracking for Gesture Tracking and Recognition. U.S. Patent No.10,241.581, 26 March 2019. [Google Scholar]
  6. Wang, N.; Shi, J.; Yeung, D.-Y.; Jia, J. Understanding and diagnosing visual tracking systems. In Proceedings of the IEEE International Conference on Computer Vision, Seoul, Korea, 27 October–3 November 2019; pp. 3101–3109. [Google Scholar]
  7. Li, K.; He, F.-Z.; Yu, H.-P. Robust visual tracking based on convolutional features with illumination and occlusion handing. J. Comput. Sci. Technol. 2018, 33, 223–236. [Google Scholar] [CrossRef]
  8. Du, X.; Clancy, N.; Arya, S.; Hanna, G.B.; Kelly, J.; Elson, D.S.; Stoyanov, D. Robust surface tracking combining features, intensity and illumination compensation. Int. J. Comput. Assist. Radiol. Surg. 2015, 10, 1915–1926. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  9. Alismail, H.; Browning, B.; Lucey, S. Robust tracking in low light and sudden illumination changes. In Proceedings of the 2016 Fourth International Conference on 3D Vision (3DV), Stanford, CA, USA, 25–28 October 2016; pp. 389–398. [Google Scholar]
  10. Kiani Galoogahi, H.; Fagg, A.; Lucey, S. Learning background-aware correlation filters for visual tracking. In Proceedings of the IEEE International Conference on Computer Vision, Seoul, Korea, 27 October–3 November 2019; pp. 1135–1143. [Google Scholar]
  11. Zhu, Y.; Mottaghi, R.; Kolve, E.; Lim, J.J.; Gupta, A.; Fei-Fei, L.; Farhadi, A. Target-driven visual navigation in indoor scenes using deep reinforcement learning. In Proceedings of the 2017 IEEE international conference on robotics and automation (ICRA), Singapore, Singapore, 29 May–3 June 2017; pp. 3357–3364. [Google Scholar]
  12. Zhao, R.; Ouyang, W.; Li, H.; Wang, X. Saliency detection by multi-context deep learning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2019; pp. 1265–1274. [Google Scholar]
  13. Higgins, I.; Matthey, L.; Glorot, X.; Pal, A.; Uria, B.; Blundell, C.; Mohamed, S.; Lerchner, A. Early visual concept learning with unsupervised deep learning. arXiv 2016, preprint. arXiv:1606.05579. [Google Scholar]
  14. Guo, Y.; Liu, Y.; Oerlemans, A.; Lao, S.; Wu, S.; Lew, M.S. Deep learning for visual understanding: A review. Neurocomputing 2016, 187, 27–48. [Google Scholar] [CrossRef]
  15. Gulshan, V.; Peng, L.; Coram, M.; Stumpe, M.C.; Wu, D.; Narayanaswamy, A.; Venugopalan, S.; Widner, K.; Madams, T.; Cuadros, J. Development and validation of a deep learning algorithm for detection of diabetic retinopathy in retinal fundus photographs. JAMA 2016, 316, 2402–2410. [Google Scholar] [CrossRef] [PubMed]
  16. Shin, H.-C.; Roth, H.R.; Gao, M.; Lu, L.; Xu, Z.; Nogues, I.; Yao, J.; Mollura, D.; Summers, R.M. Deep convolutional neural networks for computer-aided detection: CNN architectures, dataset characteristics and transfer learning. IEEE Trans. Med. Imaging 2016, 35, 1285–1298. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  17. Bertinetto, L.; Valmadre, J.; Henriques, J.F.; Vedaldi, A.; Torr, P.H. Fully-convolutional siamese networks for object tracking. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 8–16 October 2016; pp. 850–865. [Google Scholar]
  18. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. In Proceedings of the Advances in Neural Information Processing Systems, Harrahs and Harveys, Lake Tahoe, 3–8 December 2012; pp. 1097–1105. [Google Scholar]
  19. Bolme, D.S.; Beveridge, J.R.; Draper, B.A.; Lui, Y.M. Visual object tracking using adaptive correlation filters. In Proceedings of the 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, San Francisco, CA, USA, 13–18 June 2010; pp. 2544–2550. [Google Scholar]
  20. Henriques, J.F.; Caseiro, R.; Martins, P.; Batista, J. Exploiting the circulant structure of tracking-by-detection with kernels. In European Conference on Computer Vision; Springer: Berlin/Heidelberg, Germany, 2012; pp. 702–715. [Google Scholar]
  21. Danelljan, M.; Häger, G.; Khan, F.S.; Felsberg, M. Coloring channel representations for visual tracking. In Scandinavian Conference on Image Analysis; Springer: Copenhagen, Denmark, 2015; pp. 117–129. [Google Scholar]
  22. Henriques, J.F.; Caseiro, R.; Martins, P.; Batista, J. High-speed tracking with kernelized correlation filters. IEEE Trans. Pattern Anal. Mach. Intell. 2014, 37, 583–596. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  23. Ma, C.; Yang, X.; Zhang, C.; Yang, M.-H. Long-term correlation tracking. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 5388–5396. [Google Scholar]
  24. Danelljan, M.; Häger, G.; Khan, F.; Felsberg, M. Accurate scale estimation for robust visual tracking. In Proceedings of the British Machine Vision Conference, Nottingham, UK, 1–5 September 2014. [Google Scholar]
  25. Bertinetto, L.; Valmadre, J.; Golodetz, S.; Miksik, O.; Torr, P.H. Staple: Complementary learners for real-time tracking. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 1401–1409. [Google Scholar]
  26. Huang, G.; Ji, H.; Zhang, W. Adaptive spatially regularized correlation filter tracking via level set segmentation. J. Electron. Imaging 2019, 28, 063013. [Google Scholar] [CrossRef]
  27. Danelljan, M.; Hager, G.; Shahbaz Khan, F.; Felsberg, M. Learning spatially regularized correlation filters for visual tracking. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 4310–4318. [Google Scholar]
  28. Li, F.; Tian, C.; Zuo, W.; Zhang, L.; Yang, M.-H. Learning spatial-temporal regularized correlation filters for visual tracking. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 4904–4913. [Google Scholar]
  29. Qi, Y.; Zhang, S.; Qin, L.; Yao, H.; Huang, Q.; Lim, J.; Yang, M.-H. Hedged deep tracking. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 4303–4311. [Google Scholar]
  30. Ma, C.; Huang, J.-B.; Yang, X.; Yang, M.-H. Hierarchical convolutional features for visual tracking. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 3074–3082. [Google Scholar]
  31. Valmadre, J.; Bertinetto, L.; Henriques, J.; Vedaldi, A.; Torr, P.H. End-to-end representation learning for correlation filter based tracking. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2805–2813. [Google Scholar]
  32. Guo, Q.; Feng, W.; Zhou, C.; Huang, R.; Wan, L.; Wang, S. Learning dynamic siamese network for visual object tracking. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 1763–1771. [Google Scholar]
  33. Li, B.; Yan, J.; Wu, W.; Zhu, Z.; Hu, X. High performance visual tracking with siamese region proposal network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 8971–8980. [Google Scholar]
  34. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster r-cnn: Towards real-time object detection with region proposal networks. In Proceedings of the Advances in Neural Information Processing Systems, Montreal QC, Canada, 7–12 December 2015; pp. 91–99. [Google Scholar]
  35. Zhu, Z.; Wang, Q.; Li, B.; Wu, W.; Yan, J.; Hu, W. Distractor-aware siamese networks for visual object tracking. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 101–117. [Google Scholar]
  36. Dong, X.; Shen, J. Triplet loss in siamese network for object tracking. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 459–474. [Google Scholar]
  37. Zhang, Z.; Peng, H. Deeper and wider siamese networks for real-time visual tracking. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2019; pp. 4591–4600. [Google Scholar]
  38. Fan, H.; Ling, H. Siamese cascaded region proposal networks for real-time visual tracking. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2019; pp. 7952–7961. [Google Scholar]
  39. Wu, Y.; Lim, J.; Yang, M.-H. Online object tracking: A benchmark. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Portland, OH, USA, 23–28 June 2018; pp. 2411–2418. [Google Scholar]
  40. Liang, P.; Blasch, E.; Ling, H. Encoding color information for visual tracking: Algorithms and benchmark. IEEE Trans. Image Process. 2015, 24, 5630–5644. [Google Scholar] [CrossRef] [PubMed]
  41. Nam, H.; Han, B. Learning multi-domain convolutional neural networks for visual tracking. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 4293–4302. [Google Scholar]
  42. Danelljan, M.; Bhat, G.; Shahbaz Khan, F.; Felsberg, M. Eco: Efficient convolution operators for tracking. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 6638–6646. [Google Scholar]
  43. Choi, J.; Jin Chang, H.; Fischer, T.; Yun, S.; Lee, K.; Jeong, J.; Demiris, Y.; Young Choi, J. Context-aware deep feature compression for high-speed visual tracking. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 479–488. [Google Scholar]
  44. Wang, Q.; Gao, J.; Xing, J.; Zhang, M.; Hu, W. Dcfnet: Discriminant correlation filters network for visual tracking. arXiv 2017, preprint. arXiv:1704.0405. [Google Scholar]
  45. Wang, M.; Liu, Y.; Huang, Z. Large margin object tracking with circulant feature maps. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 4021–4029. [Google Scholar]
  46. Zhang, T.; Xu, C.; Yang, M.-H. Multi-task correlation particle filter for robust object tracking. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 4335–4343. [Google Scholar]
  47. Zhang, K.; Liu, Q.; Wu, Y.; Yang, M.-H. Robust visual tracking via convolutional networks without training. IEEE Trans. Image Process. 2016, 25, 1779–1792. [Google Scholar] [CrossRef] [PubMed]
  48. Danelljan, M.; Hager, G.; Shahbaz Khan, F.; Felsberg, M. Adaptive decontamination of the training set: A unified formulation for discriminative visual tracking. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 1430–1438. [Google Scholar]
  49. Danelljan, M.; Hager, G.; Shahbaz Khan, F.; Felsberg, M. Convolutional features for correlation filter based visual tracking. In Proceedings of the IEEE International Conference on Computer Vision Workshops, Santiago, Chile, 7–13 December 2015; pp. 58–66. [Google Scholar]
Figure 1. Basic workflow of our Histogram–Kalman SiamFC (HKSiamFC) tracker. The two-branch convolutional neural network (CNN) has the same architecture as the original fully-convolutional Siamese (SiamFC) tracker. We adopted the histogram model from Staple in HKSiamFC and used it to generate a histogram-score map, and we then optimized it with a Gaussian filter; this map is used to provide prior information about target appearance. We also designed a Kalman model to generate a Gaussian likelihood map based on a Kalman filter to provide prior information about target motion. Finally, we combined all three maps to capture the target.
Figure 1. Basic workflow of our Histogram–Kalman SiamFC (HKSiamFC) tracker. The two-branch convolutional neural network (CNN) has the same architecture as the original fully-convolutional Siamese (SiamFC) tracker. We adopted the histogram model from Staple in HKSiamFC and used it to generate a histogram-score map, and we then optimized it with a Gaussian filter; this map is used to provide prior information about target appearance. We also designed a Kalman model to generate a Gaussian likelihood map based on a Kalman filter to provide prior information about target motion. Finally, we combined all three maps to capture the target.
Sensors 20 02137 g001
Figure 2. (top) Four frames (#0010, #0011, #0023, and #24) of SiamFC’s tracking failure in Human8 sequence; (bottom) histogram maps of each frame corresponding to first row (#0010, #0011, #0023, and #24), the red region in the map represents high score.
Figure 2. (top) Four frames (#0010, #0011, #0023, and #24) of SiamFC’s tracking failure in Human8 sequence; (bottom) histogram maps of each frame corresponding to first row (#0010, #0011, #0023, and #24), the red region in the map represents high score.
Sensors 20 02137 g002
Figure 3. Effect of processing original histogram-score map by Gaussian filter.
Figure 3. Effect of processing original histogram-score map by Gaussian filter.
Sensors 20 02137 g003
Figure 4. Workflow of using Kalman filter to generate Gaussian likelihood maps. The green rectangle in Frame1 is the ground-truth bounding box containing the right target.
Figure 4. Workflow of using Kalman filter to generate Gaussian likelihood maps. The green rectangle in Frame1 is the ground-truth bounding box containing the right target.
Sensors 20 02137 g004
Figure 5. Comparison of HKSiamFC and five baseline trackers on Online Object Tracking Benchmark (OTB) dataset. Three plot pairs are results of (topbottom) OTB100, OTB50, and OTB2013. This picture is best viewed on high-resolution displays.
Figure 5. Comparison of HKSiamFC and five baseline trackers on Online Object Tracking Benchmark (OTB) dataset. Three plot pairs are results of (topbottom) OTB100, OTB50, and OTB2013. This picture is best viewed on high-resolution displays.
Sensors 20 02137 g005
Figure 6. Comparison of HKSiamFC and five baseline trackers using precision-plot metric under the 11 tracking scenarios. This picture is best viewed on high-resolution displays.
Figure 6. Comparison of HKSiamFC and five baseline trackers using precision-plot metric under the 11 tracking scenarios. This picture is best viewed on high-resolution displays.
Sensors 20 02137 g006
Figure 7. Comparison of HKSiamFC and five baseline trackers using success-plot metric under the 11 tracking scenarios. This picture is best viewed on high-resolution displays.
Figure 7. Comparison of HKSiamFC and five baseline trackers using success-plot metric under the 11 tracking scenarios. This picture is best viewed on high-resolution displays.
Sensors 20 02137 g007
Figure 8. Comparison between HKSiamFC and 14 state-of-the-art trackers on OTB100. This picture is best viewed on high-resolution displays.
Figure 8. Comparison between HKSiamFC and 14 state-of-the-art trackers on OTB100. This picture is best viewed on high-resolution displays.
Sensors 20 02137 g008
Figure 9. Tracking result of 15 compared trackers in precision plot in 11 different scenarios. This picture is best viewed on high-resolution displays.
Figure 9. Tracking result of 15 compared trackers in precision plot in 11 different scenarios. This picture is best viewed on high-resolution displays.
Sensors 20 02137 g009
Figure 10. Tracking result of 18 compared trackers on Temple Color (TC128). This picture is best viewed on high-resolution displays.
Figure 10. Tracking result of 18 compared trackers on Temple Color (TC128). This picture is best viewed on high-resolution displays.
Sensors 20 02137 g010
Figure 11. Qualitative tracking results of HKSiamFC and 10 state-of-the-art trackers on several typical sequences of OTB. Sequences (top–bottom) rows are Biker, Diving, Human 3, Jump, skating 2-1, and soccer. The color of each tracker is listed at the bottom of this figure.
Figure 11. Qualitative tracking results of HKSiamFC and 10 state-of-the-art trackers on several typical sequences of OTB. Sequences (top–bottom) rows are Biker, Diving, Human 3, Jump, skating 2-1, and soccer. The color of each tracker is listed at the bottom of this figure.
Sensors 20 02137 g011
Table 1. Six trackers’ average precision values and average Area Under Curve (AUC) values in 11 different scenarios.
Table 1. Six trackers’ average precision values and average Area Under Curve (AUC) values in 11 different scenarios.
OCCLRIVMBOPROVBCDEFFMIPRSV
PrecisionHKSiamFC0.8390.8170.8510.7660.8500.7050.8660.8690. 7790. 8560.825
SiamFC0.7310.8050.7730.6950.7660.6220.7910. 7060. 7090. 7990.748
Staple0.7160.6310.7820.6980.7260.6610.7660. 7830. 6890. 7660.715
HSiamFC0.7610.7720.7360.7040.7640.6240.7480. 7910. 7050. 7860.738
KSiamFC0.7280.7620.7790.6690.7520.6510.8070. 7080. 6780. 7790.721
HKSiamFC (No-Template-Update)0.5990.4620.5440.5400.6440.4350.6160. 6670. 5870. 6620.691
AUCHKSiamFC0.5990.5110.6030.5900.6080.5130.6170.5920.5790.6100.579
SiamFC0.5430.5280.5660.5540.5570.4550.5810.5030.5450.5850.540
Staple0.5420.4180.5930.5390.5270.4810.5740.5480.5320.5480.518
HSiamFC0.5560.4820.5430.5530.5530.4600.5440.5460.5360.5730.532
KSiamFC0.5400.5080.5690.5390.5520.4840.5940.5040.5320.5740.531
HKSiamFC (No-Template-Update)0.4580.3060.4030.4700.4810.3850.4640.4780.4700.4850.466
Table 2. HKSiamFC’s precision ranking in each scenario of all 15 trackers.
Table 2. HKSiamFC’s precision ranking in each scenario of all 15 trackers.
FMBCMBDEFIVIPRLROCCOPROVSV
Precision Ranking62523252282
Table 3. Precision and Area Under Curve (AUC) value of the HKSiamFC and eight other trackers on the TC128 dataset(The first and second best scores are marked with red and green colors, respectively).
Table 3. Precision and Area Under Curve (AUC) value of the HKSiamFC and eight other trackers on the TC128 dataset(The first and second best scores are marked with red and green colors, respectively).
HKSiamFCMCPFSRDCFDeep SRDCFStapleBACFSRDCFdeconHDTCNT
Precision77.876.969.674.066.866.072.968.644.9
AUC55.155.251.654.150.949.654.348.033.5

Share and Cite

MDPI and ACS Style

Li, C.; Xing, Q.; Ma, Z. HKSiamFC: Visual-Tracking Framework Using Prior Information Provided by Staple and Kalman Filter. Sensors 2020, 20, 2137. https://doi.org/10.3390/s20072137

AMA Style

Li C, Xing Q, Ma Z. HKSiamFC: Visual-Tracking Framework Using Prior Information Provided by Staple and Kalman Filter. Sensors. 2020; 20(7):2137. https://doi.org/10.3390/s20072137

Chicago/Turabian Style

Li, Chenpu, Qianjian Xing, and Zhenguo Ma. 2020. "HKSiamFC: Visual-Tracking Framework Using Prior Information Provided by Staple and Kalman Filter" Sensors 20, no. 7: 2137. https://doi.org/10.3390/s20072137

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop