Next Article in Journal
Towards a Cascading Reasoning Framework to Support Responsive Ambient-Intelligent Healthcare Interventions
Previous Article in Journal
Partial Discharge Recognition with a Multi-Resolution Convolutional Neural Network
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Visual Object Tracking Using Structured Sparse PCA-Based Appearance Representation and Online Learning

1
National Institute for Mathematical Science, 70 Yuseong-daero 1689 beon-gil, Yuseong-gu, Daejeon 34047, Korea
2
Artificial Intelligence Research Institute, 22, Daewangpangyo-ro 712beon-gil, Bundang-gu, Seongnam-si 463400, Gyeonggi-do, Korea
3
College of Computer Science, Kookmin University, 77 Jeongneung-ro, Seongbuk-gu, Seoul 02707, Korea
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Sensors 2018, 18(10), 3513; https://doi.org/10.3390/s18103513
Submission received: 19 July 2018 / Revised: 11 October 2018 / Accepted: 15 October 2018 / Published: 18 October 2018
(This article belongs to the Special Issue Audio–Visual Sensor Fusion Strategies for Video Content Analytics)

Abstract

:
Visual object tracking is a fundamental research area in the field of computer vision and pattern recognition because it can be utilized by various intelligent systems. However, visual object tracking faces various challenging issues because tracking is influenced by illumination change, pose change, partial occlusion and background clutter. Sparse representation-based appearance modeling and dictionary learning that optimize tracking history have been proposed as one possible solution to overcome the problems of visual object tracking. However, there are limitations in representing high dimensional descriptors using the standard sparse representation approach. Therefore, this study proposes a structured sparse principal component analysis to represent the complex appearance descriptors of the target object effectively with a linear combination of a small number of elementary atoms chosen from an over-complete dictionary. Using an online dictionary for learning and updating by selecting similar dictionaries that have high probability makes it possible to track the target object in a variety of environments. Qualitative and quantitative experimental results, including comparison to the current state of the art visual object tracking algorithms, validate that the proposed tracking algorithm performs favorably with changes in the target object and environment for benchmark video sequences.

1. Introduction

Visual object tracking systems have gained continuous attention and focus in the area of computer vision and pattern recognition because they can be applied to various fields, such as robotics, video surveillance, user-centered interaction systems, video communication and compression and augmented reality [1,2,3,4]. A large number of tracking algorithms has been proposed to follow the moving object in a given image sequence, while simultaneously keeping track of target identities through the significant pose changes, illumination variations and occlusions by focusing on finding appearance and motion models. To evaluate the performance of the state of the art visual object tracking methodologies quantitatively and qualitatively, benchmark tests [5,6] were conducted using a large database including ground-truth object positions to understand how these algorithms perform and effectively analyze algorithm advances.
Most state of the art visual object tracking algorithms with reported benchmark tests were formulated using the Bayesian framework [7] where the maximum a posteriori (MAP) state of the given observation was estimated by decomposing the visual object tracking system into three components.
  • An appearance model that captures the visual characteristics of the target object and evaluates the similarity between observed samples and the model.
  • A motion model that locates the target between successive frames utilizing certain motion hypotheses.
  • An optimization strategy that associates the appearance model with the motion model and finds the most likely location in the current frame.
In the Bayesian visual object tracking framework, the main issue of robust target object tracking is to find models for status and observation, such as target representation and localization, as well as filtering and data association. Target object representation and localization methodologies follow a bottom-up process that provides a variety of tools for identifying the moving object. The specific strategy for successfully locating and tracking the target object depends on features in the color, appearance and time spaces. Filtering and data association are mostly top-down processes, incorporating prior information about the scene or object, dealing with object dynamics and evaluating different hypotheses.
The core technique of visual object tracking in the Bayesian framework aims to robustly estimate the motion state of a target object with a defined appearance model in each frame from given image sequences. To achieve visual object tracking, it is necessary to categorize the appearance model into several task-specific categories. Popular appearance models used in object tracking can be separated into global and local visual appearance models [8]. Global visual representation of the target object is simple and computationally efficient for fast object tracking, but is very sensitive to target deformation and environmental changes, including illumination. A multi-cue strategy is adopted in relation to the global features, incorporating multiple visual information types, to deal with complicated appearance changes. In contrast, local visual appearance representation is robust to global appearance change by capturing the local structural object appearance. However, the representation often suffers from noise distribution and background distraction.
Sparse representation and dictionary learning for online appearance modeling have been recently proposed as an alternative solution, formulating the over-complete dictionary as a linear combination of basis functions. However, global linear sparse representation has problems with partial occlusion and local deformation. Since the dictionary uniformly emphasizes the object, occlusion and local deformation can be seen as noise when estimating similarity [9,10,11]. Another characteristic inherent in natural images is their high dimensionality, which causes complex and expensive computation. Exploration of the specific structure of sparsity as a prior enables dictionary learning to reduce computational costs effectively [12,13,14]. Therefore, we propose a structured sparse principal component analysis (PCA)-based subspace representation to represent the appearance model of the target object effectively and online learning techniques for robust visual object tracking. We use the structured sparse PCA to find a sparse linear combination over a basis library containing target and trivial templates by reducing the data dimension. The proposed structured sparse PCA-based visual object tracking within the Bayesian framework is decomposed into initialization, observation model, motion tracking model and update. The structured spare PCA-based appearance model representation and learning of domain-specific over-complete dictionaries are used to obtain MAP dictionary estimates within an appropriately chosen dictionary. The main contributions of our proposed robust visual object tracking system are as follows.
  • Structured sparse PCA-based appearance representation and learning for efficient description of the target object with few dictionary entries, to reduce the high-dimensional descriptor and to retain the structure.
  • Local structure enforced similarity measures to avoid problems from partial occlusion, illumination and background clutter.
  • Training image selection for robust online dictionary learning and updating by considering the probability that the training image contains the target, as opposed to the existing methods that choose the most recent training images.
Section 2 reviews relevant previous visual object tracking approaches, and Section 3 details tracking target objects from a given image by modeling the observation and motion using the proposed structured sparse PCA-based representation within the Bayesian framework. Section 4 quantitatively and qualitatively compares the proposed and current state of the art approaches experimentally. Section 5 summarizes the outcomes, concludes the paper and discusses future work.

2. Review of Previous Related Work

There is a rich literature in visual object tracking methodologies dealing with target object representations, search mechanisms and model updating. Sparse representation and modeling also have a fruitful literature exploiting prior information within the predefined structure of the basis library and contiguous spatial distribution of deformable target objects. We review some of the important milestones in terms of visual object tracking and sparse representation-based modeling.

2.1. Visual Object Tracking System

Many tracking methods have been proposed, largely separated into generative and deterministic methods. Generative visual object tracking methods search for the most similar region to the target object within a neighborhood, whereas discriminative methods treat tracking as a binary classification problem and aim to design a classifier to distinguish the target object from the background [15].
Early visual object tracking systems focused on generative methods, such as the Lucas–Kanade tracker [16], Kalman filter [17,18] and mean-shift (MS) tracker [19,20]. The Kalman filter [17] used for visual object tracking commonly uses the state and observation model uncertainties to calculate actual Gaussian noise, which causes certain parameter estimations to produce errors in the model, with consequent decreased estimation precision. The particle filter (PF) is efficient for conventional tracking problems with non-Gaussian distributions and multi-modality [21]. MS-based approaches are efficient for tracking non-rigid objects whose appearances are defined by histograms, but this makes them poor at dealing with illumination and/or pose variations [19,20].
Multiple instance learning (MIL)-based tracking [22] implements discriminative tracking by building a boosting classifier that tracks bags of image patches by incrementally updating the training patches over time. Online appearance learning (OAL)-based visual object tracking uses different target object appearances as a set of probability mass functions to adaptively deal with pose variations [23]. Many approaches attempted to efficiently represent the variation of rigid or limited deformation motion using an adaptive appearance model, such as incremental visual [24] and fragment-based (Frag) [25] trackers. Kelal et al. [26] proposed a paradigm for training a binary classifier from labeled and unlabeled examples called P-N learning for visual object tracking. Tracking-learning-detection (TLD) is an award-winning, real-time algorithm for tracking unknown objects in video streams that simultaneously tracks the object, learns its appearance and detects it whenever it appears in the video [27]. Struct [28] is an extended version of TLD using kernels. On the other hand, sparse representation-based visual object tracking systems like sparse collaborative appearance (SCM) [29], visual tracking decomposition (VTD) [30], the sparse representation-based l 1 tracker [31], the structured sparse tracking (SST) [32] model and sparse mask models [33,34] use an appearance model to find the sparsest linear combination of basis functions from an over-complete dictionary. However, most dictionary learning-based systems still have problems in high-dimensional reduction. Deep learning-based machine learning techniques have been recently applied to separate target objects from target candidate image templates [35,36,37,38,39] and showed a good performance to track the target object, but this requires numerous training templates.
In contrast to visual tracking approaches based on pixel-based observation models, superpixel tracking (SPT) [40] uses middle level features to both remove noise and enforce the target object color of the candidate template.

2.2. Sparse Representation-Based Learning

Sparse signal representation is an extremely powerful tool for acquiring, representing and compressing high dimensional signals. Mathematically, solving a sparse representation and learning involves seeking the sparsest linear combination of basis functions from an over-complete dictionary. The basic concept of how to represent or reconstruct signals with sparse samples is an extremely important problem in many practical fields, such as signal processing, machine learning, computer vision and robotics. Compressive sensing (CS) is based on the principle that signal sparsity can be exploited to recover the original signal from significantly less samples than required by the Shannon–Nyquist theorem [41,42]. Generally, CS algorithms include three basic components: sparse representation, encoding measuring and a reconstruction [12]. In particular, sparse representation that approximately solves a system of equations with sparse vectors is popularly applied for pattern recognition because it exploits a linear combination of training samples to represent the test sample and computes sparse representation coefficients of the linear representation system [43,44,45].
Structured sparse representation is an extension of standard sparse representation in statistical signal processing and learning [46,47]. Motivated by potential group structures on feature sets, group sparse representation has become popular in recent years. Group sparsity is used not only for estimating hyper-parameters in the sparse prior model, but also for group least absolute shrinkage and selection operator (LASSO). Techniques using strong group for group LASSO have been developed and show superior performance for strongly group-sparse feature sets [48]. However, group LASSO works well only under the strong group sparsity assumption and does not apply for more general structures, such as overlapping groups, and tonal or transient structures. Therefore, Huang et al. [14] proposed that sparse representation can be solved by a structured greedy algorithm when a coding scheme can be approximated by block coding with base blocks.

3. Structured Sparse PCA-Based Tracking and Online Dictionary Learning

For visual object tracking, it is reasonable to assume that the object trajectory is continuous and object features are consistent or change insignificantly over a short time interval. Thus, once a representation of the feature vector is found in terms of fix-ahead dictionaries, consecutive representations of the feature vectors are almost constant. Therefore, we propose an object tracking method by classifying the target appearance model’s coefficients. The dictionaries are generated from appearance features by applying structured sparse PCA and updated using the last data. The object tracking comprises three modes: observation, tracking and update within the Bayesian framework, as shown in Figure 1.

3.1. Notations and Symbols

Before proceeding to the technical details, we introduce the notations and symbols used throughout this paper, as shown in Table 1. Lower case letters denote real variables, and upper case (capital) letters denote multi-dimensional variables, such as images and matrices, except for the case Y t , which denotes an observation random variable taking real numbers. Column vectors given are shown as boldface, and mappings are denoted by letters of the Greek alphabet.

3.2. Bayesian Framework-Based Visual Object Tracking

The traditional visual object tracking algorithm can be formulated with the Bayesian framework where the maximum a posteriori (MAP) estimation of the state given the observations up to time t is expressed as:
p ( X t | Y 1 : t ) = p ( Y t | Y 1 : t 1 , X t ) p ( X t | Y 1 : t 1 ) p ( Y t | Y 1 : t 1 ) = p ( Y t | X t ) n t p ( X t | X t 1 ) p ( X t 1 | Y 1 : t 1 ) d X t 1 ,
where X t is the state at t; Y 1 : t denotes all the observations up to t; and n t is a normalization term,
n t = p ( Y t | Y 1 : t 1 ) = p ( X t | Y 1 : t 1 ) p ( Y t | Y 1 : t 1 , X t ) d X t .
We use the following assumptions.
(i)
State X t is independent of the past given the present X t 1 ,
p ( X t | X 1 : t 1 , Y 1 : t 1 ) = p ( X t | X t 1 ) .
(ii)
Observations Y 1 : t are conditionally independent given X t ,
p ( Y t | Y 1 : t 1 , X t ) = p ( Y t | X t ) .
We also employed the Chapman–Kolmogorov equation for Equation (1),
p ( X t | Y 1 : t 1 ) = p ( X t | X t 1 ) p ( X t 1 | Y 1 : t 1 ) d X t 1 .
In the visual object tracking scheme, the target state is defined as X t = ( x t c , w t s x , h t s y ) , where x t c represents the center location of the target and w t s x and h t s y denote its scale in the x and y directions, respectively. In terms of observation, we need to construct an effective observation model p ( Y t | X t ) and an efficient motion model p ( X t | X t 1 ) . The state estimate of the target X t at time t can be obtained by the MAP estimate over the M samples X t j and its measurements Y t j for j = 1 , , M , given X t 1 ,
X t = argmax X t j p ( X t j | Y t j , X t 1 ) .
It is worth noting that even though we need the measurement quantities p ( Y t j | X t 1 ) in solving the optimization (5) from Bayes’ rule p ( x | y , z ) = p ( y | x , z ) p ( x | z ) / p ( y | z ) , we may regard the denominator as a constant for all j = 1 , , M and solve the maximization by finding the maximum of likelihood times prior as given in (13). This is because given X t , the measurements (evidence) Y t + 1 and Y t for the two consecutive targets X t + 1 and X t remain the same. We shall see this precisely in Section 3.4.
Based on the MAP estimation, we decompose the visual object tracking procedure into:
  • structured sparse PCA-based observation and appearance representation using deterministic target object separation from background patch images,
  • motion tracking and
  • online update.

3.3. Deterministic Modeling Using Structured Sparse PCA-Based Appearance Representation

To construct the dictionary from the t 0 initial image sequences, we extract image patches using windows surrounding the target object for each t = 1 , , t 0 . Figure 1 shows the proposed procedure to separate the target object and background image patches around the target object, representing appearances using structured sparse representation. Let us explain the learning mode of the target object tracking in more detail. We create tracking dictionary vectors { d i } i = 1 r by applying feature descriptors extracted from observation frames I 1 : t 0 to the structured sparse PCA algorithm as follows.
  • We take the same sized image patches { p t t a r g e t } t = 1 t 0 centered at ( x t c , y t c ) from frames I 1 : t 0 , respectively.
    Recall that states X t = ( x t c , w t s x , h t s y ) consist of the center location x t c = ( x t x , y t c ) of the target and its window size ( w t s x , h t s y ) in the x and y directions, respectively. From each patch p t t a r g e t , t = 1 , , t 0 , we construct the descriptor v t t g R s of the target object by sequentially accumulating gradient histograms from equally-divided subregions of p t t a r g e t .
    To enhance tracking performance, we also create background feature descriptors v j b g R s from the four background patch images { p t , ( a x , b y ) b a c k I t | a x , b y = 1 , 1 and a x 2 + b y 2 = 1 , t = 1 , , t 0 } around the target patch p t t a r g e t as follows.
    • For each t = 1 , , t 0 , patches p t , ( a x , b y ) b a c k are subimages of I t centered at ( x t c + a x w t s x , y t c + b y h t s y ) with the same size as p t t a r g e t .
    • When the domain of p t , ( a x , b y ) b a c k does not entirely belong to that of I t , we regard it as an empty set.
    • Let { v j b g } j = 1 κ R s with κ 4 t 0 be background appearance descriptors obtained from background patches p t , ( a x , b y ) b a c k in the same manner used to create the target descriptors.
  • After creating the appearance feature descriptors v t t g and v j b g , we apply the constrained structured sparse PCA dictionary learning algorithm to the target and background descriptors to find dictionaries { d i } i = 1 r R s ,
    ( D , C ) = argmin D R s × r C R r × ( t 0 + κ ) H ( D , C ) subject to c j 2 1 , j = 1 , t 0 + κ ,
    where the objective function H ( D , C ) is given by:
    H ( D , C ) = 1 2 s ( t 0 + κ ) V D C F 2 + λ i = 1 r Ω ν ( d i )
    and V = ( v i ) i = 1 t 0 + κ is the s × ( t 0 + κ ) matrix with v 1 : t 0 t g and v 1 : κ b g column vectors; D = ( d i ) i = 1 r R s × r is the dictionary matrix; and C = ( c i ) i = 1 t 0 + κ R r × ( t 0 + κ ) is the coefficient matrix, such that for i = 1 , , t 0 + κ , v i is (approximately or exactly) expressed by a linear combination of d j with coefficients c i = ( c j i ) j = 1 r ,
    v i j = 1 r c j i d j = D c i , c i = ( c 1 i , c 2 i , , c r i ) T
    for i = 1 , , t 0 + κ .
  • Let · F be the Frobenius matrix norm, A F 2 = t r a c e ( A A T ) = i = 1 n j = 1 m a i j 2 , for A = ( a i j ) R n × m ; · 2 the Euclidean norm; and Ω ν a quasi-norm that controls the sparsity and structure of the support of d j . In this work, the quasi-norm Ω ν is defined as follows. Let G 1 , G 2 , G 3 , G 4 be four mutually disjoint subsets of { 1 , 2 , , s } . Then, every vector d = ( d 1 , , d s ) R s is decomposed into four subvectors d k = ( d 1 k , , d s k ) , k = 1 , 2 , 3 , 4 such that for 1 k 4 and 1 j s ,
    d j k = d j , if   j G k 0 , otherwise
    Then, Ω ν ( d ) is defined as:
    Ω ν ( d ) = d 1 2 1 2 + d 2 2 1 2 + d 3 2 1 2 + d 4 2 1 2 2 .
    We refer to [49] and the references therein for details on the quasi-norm. The decomposition of V into D C enables us to reduce the dimensionality of the descriptors using Equation (6).
    Although there is clearly a limitation in representing high dimensional descriptors using a smaller number of vectors than the dimension, the proposed structured sparse PCA is more effective to represent nonlinear and high dimensional descriptors by reducing the dimension while retaining the target object structure. For more details of structured sparse PCA algorithms, refer to the original paper [49].
  • Finally, we find a linear support vector machine (SVM) Φ : R s R , such that Φ ( ( D C ) i ) 1 ( i = 1 , , t 0 ) for the target feature-related column vectors of D C and Φ ( ( D C ) i ) 1 ( i = t 0 , , t 0 + κ ) for the background appearance feature related column vectors of D C , where ( D C ) i denotes the i-th column vector of D C , i.e., ( D C ) i = D c i . Using the classifier Φ , we estimate observation Y t { 1 , 1 } as:
    Y t = 1 ( target ) , if   Φ ( v t t g ) 0 1 ( background ) , otherwise ,
    where we recall that v t t g is the target feature descriptor obtained from state X t . Note that when the target object is occluded or not observed, the value of the observation becomes negative.
The procedure of deterministic separation using the structured sparse PCA-based representation of the target and the background is shown in Algorithm 1.
Algorithm 1: Discriminative classification of target objects.
Input: frame images I 1 : t 0 , states X 1 : t 0 , integers r t g , r b g > 0
 1.
take target patches p 1 : t 0 t a r g e t
 2.
take background patches p t , ( a x , b y ) b a c k , t = 1 , , t 0
 3.
create target appearance descriptors v 1 : t 0 t g R s
 4.
create background appearance descriptors v 1 : κ b g R s
 5.
find ( D , C ) by applying structured sparse PCA (6)
 6.
find optimized classifier Φ such that Φ ( ( D C ) i ) 1
for i = 1 , , t 0 and Φ ( ( D C ) i ) 1 for i = t 0 , , t 0 + κ
Output: target appearance descriptors v 1 : t 0 t g R s and classifier Φ

3.4. Motion Tracking Model and Online Update

Using the learned dictionary of the target object and classifier, we track the target object for frames { I t + 1 } t + 1 > t 0 from the previous states X t . The motion model p ( X t + 1 | X t ) starts from the Gaussian assumption:
p ( X t + 1 | X t ) = N ( X t + 1 ; X t , σ ) = 1 ( 2 π ) 2 | σ | 1 / 2 exp 1 2 ( X t + 1 X t ) T σ 1 ( X t + 1 X t ) ,
where σ is a diagonal covariance matrix whose elements are the standard deviations for location and size and | σ | is the determinant of σ .
Let I t + 1 be the frame at t + 1 > t 0 , and assume we already have states X 1 : t . We randomly take M candidate states { X ^ t + 1 j } j = 1 M around ( x t c , y t c ) in I t + 1 with X ^ t + 1 j N ( X t , σ ) . Similar to the observation mode, we build the M descriptors { v t + 1 j } j = 1 M from sample states { X ^ t + 1 j } j = 1 M .
Since the observation model p ( Y t | X t ) with given state X t 1 implies the confidence of an observation Y t at state X t being the target, the likelihood p ( Y t + 1 | X ^ t + 1 j , X t ) is proportional to its confidence:
p ( Y t + 1 | X ^ t + 1 j , X t ) ω ( Y t + 1 | X ^ t + 1 j , X t ) .
Given the target state X t at time t , the confidence ω ( y | X t + 1 , X t ) for the target candidates X t + 1 with positive confidence value increases as we observe the targets in a larger area, whereas confidence for target candidates with negative confidence decreases. Therefore, we evaluate confidence ω ( y | X t + 1 , X t ) comparing with state X t as:
ω ( y | X t + 1 , X t ) = 1 1 + e y Φ ( v t + 1 ) · w t + 1 s x + h t + 1 s y w t s x + h t s y y ,
where y = 1 , 1 and v t + 1 is the feature descriptor extracted from the target state X t + 1 and w t + 1 s x · h t + 1 s y denotes the window size of X t + 1 . We note that in the tracking mode, we estimate the observation in (7) and the confidence in (9) by applying the descriptor v directly to the SVM, Φ ( v ) , instead of using the dictionary representation ( D T D ) 1 D T v as we construct the SVM Φ in the initialization mode. This is because the descriptor v and its dictionary representation ( D T D ) 1 D T v are much similar for ( D T D ) 1 D T v , which minimizes v D w 2 , so that it is cheaper to apply the descriptor to SVM rather than to utilize the representation, which requires the computation of the inverse matrix ( D T D ) 1 .
Now, the likelihood p ( Y t + 1 | X ^ t + 1 j , X t ) of Y t + 1 given states X ^ t + 1 j and X t is defined as:
p ( Y t + 1 | X ^ t + 1 j , X t ) = 1 n ω ω ( Y t + 1 | X ^ t + 1 j , X t ) ,
for j = 1 , 2 , , M with the normalizing factor n ω = ω ( 1 | X ^ t + 1 j , X t ) + ω ( 1 | X ^ t + 1 j , X t ) . Applying the motion model p ( X ^ t + 1 j | X t ) obtained from Equation (8) and the observation model p ( Y t + 1 | X ^ t + 1 j , X t ) obtained from Equation (10) to the Bayesian formulation in Equation (1), we estimate the a posteriori probability p ( X ^ t + 1 j | Y t + 1 , X t ) as:
p ( X ^ t + 1 j | Y t + 1 , X t ) = p ( Y t + 1 | X ^ t + 1 j , X t ) p ( X ^ t + 1 j | X t ) p ( Y t + 1 | X t ) = p ( Y t + 1 | X ^ t + 1 j , X t ) p ( X ^ t + 1 j | X t ) 1 + e Y t + 1 Φ ( v t ) .
Finally, we obtain the most likely target state X t + 1 at t + 1 with estimated MAP over the M samples X ^ t + 1 j and its observations Y ^ t + 1 j for j = 1 , , M , given X t ,
X t + 1 = argmax X ^ t + 1 j , 1 j M p ( X ^ t + 1 j | Y ^ t + 1 j , X t ) = argmax X ^ t + 1 j , 1 j M p ( Y ^ t + 1 j | X ^ t + 1 j , X t ) p ( X ^ t + 1 j | X t ) 1 + e Y ^ t + 1 j Φ ( v t ) .
On the other hand, it is reasonable to infer that the maximizing target state X t + 1 is very similar to X t , which implies Y t + 1 Φ ( v t ) 0 , so that Y t + 1 Φ ( v t ) Y ^ t + 1 j Φ ( v t ) for all 1 j M . From this aspect, let X ˜ t + 1 be a sample state such that Y ˜ t + 1 Φ ( v t ) 0 and the solution to the maximization:
X ˜ t + 1 = argmax X ^ t + 1 j , 1 j M p ( Y ^ t + 1 j | X ^ t + 1 j , X t ) p ( X ^ t + 1 j | X t ) .
Then, for all 1 j M , we have:
p ( Y ˜ t + 1 | X ^ t + 1 , X t ) p ( X ^ t + 1 | X t ) p ( Y ^ t + 1 j | X ^ t + 1 j , X t ) p ( X ^ t + 1 j | X t ) · 1 + e Y ˜ t + 1 Φ ( v t ) 1 + e Y ˜ t + 1 Φ ( v t ) p ( Y ^ t + 1 j | X ^ t + 1 j , X t ) p ( X ^ t + 1 j | X t ) · 1 + e Y ˜ t + 1 Φ ( v t ) 1 + e Y ^ t + 1 j Φ ( v t )
for 1 + exp ( Y t + 1 Φ ( v t ) ) 1 + exp ( Y ^ t + 1 j Φ ( v t ) ) , so that:
p ( Y ˜ t + 1 | X ^ t + 1 , X t ) p ( X ^ t + 1 | X t ) 1 + e Y ˜ t + 1 Φ ( v t ) p ( Y ^ t + 1 j | X ^ t + 1 j , X t ) p ( X ^ t + 1 j | X t ) 1 + e Y ^ t + 1 j Φ ( v t )
for all j = 1 , , M . This shows that we may regard the denominator 1 + e Y ^ t + 1 Φ ( v t ) in (12) as a constant for all j = 1 , , M .
Figure 2 shows the steps of how to detect the target object when a new frame comes in. M candidate samples are separated into positive and negative labels using Φ ( v ) . Usually, the ideal target template contains all of the target features, although there is some background. However, in most cases, a sample with the highest probability tends to contain less background. Figure 3a illustrates this problem. The first row of Figure 3a shows candidate samples sorted without the window size ratio in Equation (9). The ideal candidate sample is located in the fourth. However, the second row, which applied the window size ratio in Equation (9), shows that there is the ideal candidate in the first position. Consequently, we prioritize templates with the same or similar Φ such that larger window sizes are assigned a larger weight, based on the scale information of the last target estimate X t 1 (see Equation (9)). Figure 3b illustrates how the result changes when the prioritization is applied.
The proposed motion tracking model is summarized in Algorithm 2.
Algorithm 2: Motion tracking model.
 for t + 1 = t 0 + 1 to the end of the frame sequence
 1.
take M candidate states { X ^ t + 1 j } j = 1 M N ( X t , σ )
around the point ( x t c , y t c )
 2.
build up the M descriptors { v t + 1 j } j = 1 M
and their measurements { Y ^ t + 1 j } j = 1 M
 3.
compute the motion model p ( X ^ t + 1 j | X t ) (Equation (8))
 4.
compute the observation model p ( Y ^ t + 1 j | X ^ t + 1 j , X t ) by Equation (10)
 5.
estimate the a posteriori prob. p ( X ^ t + 1 j | Y ^ t + 1 j , X t ) using Equation (11)
 6.
find the most likely target state X t + 1 by Equation (12)
 7.
create the target descriptor v t + 1 t g R s
 8.
create the background descriptors v t + 1 , ( a x , b x ) b g R s
 end
Since the appearances of the target may change during tracking, we need to update the classifier Φ every k frames by updating the dictionaries as follows.
  • We save the t 0 target descriptors v 1 : t 0 t g R s into a set F = { v 1 t g , , v t 0 t g } at time t = t 0 .
  • At every t > t 0 , if p ( X t j | Y t j , X t 1 ) > θ p , we add the target descriptor v t t g and background descriptors v t , ( a x , b x ) b g to F. Otherwise, k p = k p + 1 .
  • After every k frames, we create the dictionary matrix D w and coefficient matrix C w using the vectors in F by applying the structured sparse PCA.
  • Similar to the initiation algorithm, we update Φ using the new D w and C w .
  • We check Φ ( v ) for all target descriptors v F and sort the descriptors according to their values, while keeping the k 0 largest target descriptors in F and deleting the remaining target descriptors and all the background descriptors from F.
The update interval k = k 0 + k p is between the range of k 0 and 2 k 0 . Because occlusion frames do not have (whole or partial) target patch, we need to update the dictionary slowly by increasing the value of k 0 .
We continuously update the training dictionaries using the k 0 prior templates, which have a high probability, from the target as shown in Algorithm 3. This way, if the confidence of the target is high, it will participate in the update continuously. Therefore, the target models with high confidence in the previous update and the target models in recent frames participate in the update. The target models in recent frames keep tracking when the appearance of the target object is almost unchanged, and the target models with high confidence help tracking to not fail when the appearance of the target object changes suddenly. Figure 4 shows the target models in F at the update time and the detection of the changed target appearance after the update. In the 84th frame, the top k 0 target models from previous updates are different from the current target appearance, but show a similar look to the target in the 94th frame. It can be seen that this is more suitable for detecting the changed appearance.
Algorithm 3: Dictionary update.
  for t = t 0 + 1 to the end of the frame sequence
   1. if p ( X t j | Y t , X t 1 ) > θ p
    1-1. add the target descriptor v t t g
    and background descriptors v t , ( a x , b x ) b g to F
   1. else
    1-2. k p = k p + 1
  for every k frames
   2. build the new metrics D w and C w using the vectors in F
   by structured sparse PCA
   3. update classifier Φ using D w and C w
   4. compute Φ ( v ) for v F
   5. keep the k 0 largest target descriptors in F,
   and delete the rest descriptor from F
   6. k p = 0
  end

4. Experimental Validation

This section validates the robustness of the proposed method by quantitatively and qualitatively comparing it to current state of the art approaches using the TS-50 public visual object benchmark video sequences (available online: http://cvlab.hanyang.ac.kr/tracker_benchmark/datasets.html (accessed on 8 May 2012)). The benchmark sequences include background clutter (BC), deformation (DEF), fast motion (FM), in-plane rotation (IPR), illumination variation (IV), low resolution (LR), motion blur (MB), occlusion (OCC), out-of-plane rotation (OPR), out-of view (OV) and scale variation (SV). The proposed tracker was implemented in MATLAB on a standard 4-GHz machine with 2 GB RAM. To create the descriptors, we resize all patches to [72, 72] and use the scale-invariant feature transform (SIFT) [50]. The number of samples M is set to 600. t 0 and r are set to three and 30, respectively. The k 0 and θ p are set to 10 and 0.2, respectively. We also tested the prototype VTD [30], MS [19], MIT [22], SCM [29], Frag [25], IVT [24], TLD [27], Struct [28], and ASLA [11]. The experimental results are compared in Table 2.
The proposed method can be extended to track the target object using the observation model by incorporating various descriptors, and the results are presented in the Supplementary Material. All the MATLAB code and results are available on our web site.

4.1. Qualitative Analysis

The public TS-50 video sequences used in the experiments include illumination change, partial occlusion, background clutter, low resolution and pose variations. The proposed structured sparse PCA-based visual object tracking system addresses the main problems by feature optimization and dimensionality reduction.

4.1.1. Significant Occlusion

Heavy occlusion leads to target object tracking drift due to a lack of features, but the learned local structure of the appearance model and online updating prevent the proposed tracker from creating a bias toward part of the target, mitigating the influence of background pixels. Figure 5 shows that although the target object undergoes significant occlusion for a long period, the tracker robustly retains the key appearance structure, reducing the background effect. The Girl sequence in particular shows heavy occlusion from an object with a similar shape to the target object, but the proposed system retains target tracking.

4.1.2. Illumination Change

The appearance model using structured sparse representation with an SIFT descriptor is relatively insensitive to illumination changes. Figure 6 shows that although the image sequences include significant illumination changes, the target object remains continuously within the bounding box using the proposed tracking system. Simultaneous update of target images and retention of important structures using the structure sparse PCA method ensure the proposed system continuously tracks the target object even with large illumination changes.

4.1.3. Background Clutter

Discriminative classification of the target object and background images provides clear separation between the target object and background, which have similar color, appearance and motion. Figure 7 shows that the separation of the background and target is very robust against background clutter changes.

4.2. Quantitative Analysis

We obtained the ground-truth reference values for the eight image sequences, and employed the average of overlap scores (AOS) between the tracking window and ground truth center to quantify the proposed and reference tracker performances [6]. As shown in Table 2, our proposed approach is good for deformation, fast motion, out-of-plane rotation (OPR) and out-of-view (OV), but showed balanced performance per various challenging issues in the visual object tracking. Struct [28] shows a robust performance for various performance test. SCM [29] has good performance in background clutter, illumination variation, occlusion and scale variation because it extracts the features of the target object using sparse representation, but still has variation in the video sequences like fast motion and motion blur. Figure 8 compares the performances for the proposed and current state of the art trackers for the various image sequences. The proposed tracker system tracks the target object under the partial occlusion, drift, background clutter, scale and pose variation challenges.

5. Conclusions

We proposed a structured sparse PCA-based visual object tracking incorporating initialization, motion tracking and online dictionary learning and update. In the initialization stage, a discriminative classifier was applied to target object and background image template coefficients extracted from the structured sparse PCA. The best candidate samples were selected by jointly evaluating the appearance distance and learned classifier. Online dictionary learning was based on a sparse representation appearance model where the dictionary and classifier were continuously updated. The structured sparse PCA provided dimensionality reduction of high dimensional descriptors, while retaining the structure of the appearance model.
We experimentally evaluated the effectiveness of the proposed tracking system by comparing with the twelve current state of the art trackers using eight publicly available benchmark image sequences. The proposed method performed favorably against all current trackers and was able to handle all the various tracking challenge scenarios. Quantitative and qualitative comparison of the outcomes from the challenging image sequences validated the effectiveness and robustness of the proposed algorithm.
Thus, exploiting a linear combination of key structure features using structured sparse PCA is a robust method to track target objects through illumination, partial occlusion and background clutter changes, because the structure of the appearance model effectively estimates the similarity between the target object and candidates.

Author Contributions

G.-J.Y., H.J.H., and S.M.Y. conceived and designed the experiments; H.J.H. performed the experiments; G.-J.Y. and S.M.Y. analyzed the data; G.J.Y. and S.M.Y. wrote the paper.

Funding

S.M.Y. was supported through the National Research Foundation of Korea grant funded by the Korean Ministry of Education (NRF-2016R1D1A1B04932889) and by the Korean Government (MSIP)(Grant Number 2015R1A5A7037615). G.-J.Y. was supported by National Institute for Mathematical Sciences (NIMS) founded by the Ministry of Science, ICT and Future Planning (B18220000).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Trucco, E.; Plakas, K. Video tracking: A concise survey. IEEE J. Ocean. Eng. 2006, 31, 520–529. [Google Scholar] [CrossRef]
  2. Yilmaz, A.; Javed, O.; Shah, M. Object tracking. ACM Comput. Surv. 2006, 38, 1–45. [Google Scholar] [CrossRef]
  3. Jalal, A.S.; Singh, V. The State-of-the-Art in Visual Object Tracking. Informatica 2012, 36, 227–248. [Google Scholar]
  4. Smeulders, A.W.M.; Chu, D.M.; Cucchiara, R.; Calderara, S.; Dehghan, A.; Shah, M. Visual tracking: An experimental survey. IEEE Trans. Pattern Anal. Mach. Intell. 2014, 36, 1442–1468. [Google Scholar] [PubMed]
  5. Wu, Y.; Lim, J.; Yang, M.H. Online object tracking: A benchmark. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Portland, OR, USA, 23–28 June 2013; pp. 2411–2418. [Google Scholar]
  6. Wu, Y.; Lim, J.; Yang, M.H. Object tracking benchmark. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 37, 1834–1848. [Google Scholar] [CrossRef] [PubMed]
  7. Beymer, D.; McLauchlan, P.; Coifman, B.; Malik, J. A real-time computer vision system for measuring traffic parameters. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, San Juan, Puerto Rico, 17–19 June 1997; pp. 495–501. [Google Scholar] [Green Version]
  8. Li, X.; Hu, W.; Shen, C.; Zhang, Z.; Dick, A.; van den Hengel, A. A Survey of Appearance Models in Visual Object Tracking. ACM Trans. Intell. Syst. Technol. 2013, 4, 1–42. [Google Scholar] [CrossRef]
  9. Chen, F.; Wang, Q.; Wang, S.; Zhang, W.; Xu, W. Object tracking via appearance modeling and sparse representation. Image Vis. Comput. 2011, 29, 787–796. [Google Scholar] [CrossRef]
  10. Bai, T.; Li, Y.F. Robust visual tracking with structured sparse representation appearance model. Pattern Recognit. 2012, 45, 2390–2404. [Google Scholar] [CrossRef]
  11. Jia, X.; Lu, H.; Yang, M.H. Visual tracking via adaptive structural local sparse appearance model. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Providence, RI, USA, 16–21 June 2012; pp. 1822–1829. [Google Scholar]
  12. Rubinstein, R.; Bruckstein, A.M.; Elad, M. Dictionaries for sparse representation modeling. Proc. IEEE 2010, 98, 1045–1057. [Google Scholar] [CrossRef]
  13. Sadeghi, M.; Babaie-Zadeh, M.; Jutten, C. Dictionary learning for sparse decomposition: A novel approach. IEEE Signal Process. Lett. 2013, 20, 1195–1198. [Google Scholar] [CrossRef] [Green Version]
  14. Huang, T. Linear spatial pyramid matching using sparse coding for image classification. In Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009; pp. 1794–1801. [Google Scholar]
  15. Henrigues, J.F.; Caseiro, R.; Martins, P.; Batista, J. High-Speed Tracking with Kernelized Correlation Filters. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 37, 583–596. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  16. Birchfield, S.T. KLT: An Implementation of the Kanade-Lucas-Tomasi Feature Tracker. Available online: https://cecas.clemson.edu/~stb/klt/ (accessed on 17 October 2018).
  17. Kalman, R.E. A new approach to linear filtering and prediction problems. J. Basic Eng. 1960, 82, 35–45. [Google Scholar] [CrossRef]
  18. Ramos, J.A. A kalman-tracking filter approach to nonlinear programming. Comput. Math. Appl. 1990, 19, 63–74. [Google Scholar] [CrossRef]
  19. Comaniciu, D.; Ramesh, V.; Meer, P. Real-time tracking of non-rigid objects using mean shift. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Hilton Head, SC, USA, 13–15 June 2000; Volume 2, pp. 142–149. [Google Scholar]
  20. Allen, J.G.; Xu, R.Y.D.; Jin, J.S. Object Tracking Using CamShift Algorithm and Multiple Quantized Feature Spaces. Reproduction 2006, 36, 3–7. [Google Scholar]
  21. Khan, Z.; Balch, T.; Dellaert, F. An MCMC-Based Particle Filter for Tracking Multiple Interacting Targets. In European Conference on Computer Vision; Springer: Berlin/Heidelberg, Germany, 2004; pp. 279–290. [Google Scholar]
  22. Babenko, B.; Yang, M.H.; Belongie, S.J. Visual tracking with online multiple instance learning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009; pp. 983–990. [Google Scholar]
  23. Maraghi, T.F.E.; Fleet, D.J.; Jepson, A.D. Robust online appearance models for visual tracking. In Proceedings of the the IEEE Conference on Computer Vision and Pattern Recognition, Kauai, HI, USA, 8–14 December 2001; pp. 415–422. [Google Scholar]
  24. Ross, D.A.; Lim, J.W.; Lin, R.S.; Yang, M.H. Incremental learning for robust visual tracking. Int. J. Comput. Vis. 2008, 77, 125–141. [Google Scholar] [CrossRef]
  25. Srikrishnan, V.; Nagaraj, T.; Chaudhuri, S. Fragment based tracking for scale and orientation adaptation. In Proceedings of the Indian Conference on Computer Vision, Graphics and Image Processing, Bhubaneswar, India, 16–19 December 2008; pp. 328–335. [Google Scholar]
  26. Kalal, Z.; Matas, J.; Mikolajczyk, K. P-N learning: Bootstrapping. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, San Francisco, CA, USA, 13–18 June 2010; pp. 49–56. [Google Scholar]
  27. Kalal, Z.; Mikolajczyk, K.; Matas, J. Tracking-learning-detection. IEEE Trans. Pattern Anal. Mach. Intell. 2012, 34, 1409–1422. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  28. Hare, S.; Saffari, A.; Torr, P.H.S. Struck: Structured output tracking with kernels. In Proceedings of the International Conference on Computer Vision, Barcelona, Spain, 6–13 November 2011; pp. 263–270. [Google Scholar]
  29. Zhong, W.; Lu, H.; Yang, M.H. Robust object tracking via sparse collaborative appearance model. IEEE Trans. Image Process. 2014, 23, 2356–2368. [Google Scholar] [CrossRef] [PubMed]
  30. Kwon, J.; Lee, K.M. Visual tracking decomposition. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, San Francisco, CA, USA, 13–18 June 2010; pp. 1269–1276. [Google Scholar]
  31. Bao, C.; Wu, Y.; Ling, H.; Ji, H. Real time robust L1 tracker using accelerated proximal gradient approach. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Providence, RI, USA, 16–21 June 2012; pp. 1830–1837. [Google Scholar]
  32. Zhang, T.; Liu, S.; Xu, C.; Yan, S.; Ghanem, B.; Ahuja, N.; Yang, M.-H. Structural Sparse Tracking. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; pp. 150–158. [Google Scholar]
  33. Zhang, T.; Ghanem, B.; Liu, S.; Ahuja, N. Robust Visual Tracking via Structured Multi-Task Sparse Learning. Int. J. Comput. Vis. 2017, 101, 367–383. [Google Scholar] [CrossRef]
  34. Chen, Z.; You, X.; Zhong, B.; Li, J. Dynamically Modulated Mask Sparse Tracking. IEEE Trans. Cybern. 2017, 47, 3706–3718. [Google Scholar] [CrossRef] [PubMed]
  35. Wang, N.; Yeung, D.-Y. Learning a deep compact image representation for visual tracking. In Proceedings of the Neural Information Processing Systems, Lake Tahoe, NV, USA, 5–10 December 2013; pp. 809–817. [Google Scholar]
  36. Hong, S.; You, T.; Kwak, S.; Han, B. Online tracking by learning discriminative saliency map with convolutional neural network. In Proceedings of the International Conference on Machine Learning, Lille, France, 6–11 July 2015. [Google Scholar]
  37. Zhang, D.; Maei, H.; Wang, X.; Fang, Y. Deep Reinforcement Learning for Visual Object Tracking. arXiv, 2017; arXiv:1701.08936. [Google Scholar]
  38. Nam, H.; Han, B. Learning Multi-Domain Convolutional Neural Networks for Visual Tracking. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017. [Google Scholar]
  39. Wang, L.; Ouyang, W.; Wang, X.; Lu, H. STCT: Sequentially Training Convolutional Networks for Visual Tracking. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016. [Google Scholar]
  40. Yang, F.; Lu, H.; Yang, M.-H. Robust superpixel tracking. IEEE Trans. Image Process. 2014, 23, 1639–1651. [Google Scholar] [CrossRef] [PubMed]
  41. Donoho, D.L. Compressed sensing. IEEE Trans. Inf. Theory 2006, 52, 1289–1306. [Google Scholar] [CrossRef]
  42. Candes, E.; Wakin, M. An introduction to compressive sensing. IEEE Signal Process. Mag. 2008, 25, 21–30. [Google Scholar] [CrossRef]
  43. Cheng, H. Sparse Representation, Modeling and Learning in Visual Recognition—Theory, Algorithms and Applications; Series Advances in Computer Vision and Pattern Recognition; Springer: New York, NY, USA, 2015. [Google Scholar]
  44. Kreutz-Delgado, K.; Murray, J.F.; Rao, B.D.; Engan, K.; Lee, T.-W.; Sejnowski, T.J. Dictionary learning algorithms for sparse representation. Neural Comput. 2003, 15, 349–396. [Google Scholar] [CrossRef] [PubMed]
  45. Wright, J.; Ma, Y.; Mairal, J.; Sapiro, G.; Huang, T.S.; Yan, S. Sparse representation for computer vision and pattern recognition. Proc. IEEE 2010, 98, 1031–1044. [Google Scholar] [CrossRef]
  46. Elhamifar, E.; Vidal, R. Robust classification using structured sparse representation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Colorado Springs, CO, USA, 20–25 June 2011; pp. 1873–1879. [Google Scholar]
  47. Bronstein, A.M.; Sprechmann, P.; Sapiro, G. Learning efficient structured sparse models. arXiv, 2012; arXiv:1206.4649. [Google Scholar]
  48. Varshney, K.R.; Çetin, M.J.W.; Fisher, J.W., III; Willsky, A.S. Sparse representation in structured dictionaries with application to synthetic aperture radar. IEEE Trans. Signal Process. 2008, 56, 3548–3561. [Google Scholar] [CrossRef] [Green Version]
  49. Jenatton, R.; Obozinski, G.; Bach, F.R. Structured sparse principal component analysis. In Proceedings of the Thirteenth International Conference on Artificial Intelligence and Statistics (AISTATS), Sardinia, Italy, 13–15 May 2010; pp. 66–373. [Google Scholar]
  50. Lowe, D.G. Distinctive Image Features from Scale-Invariant Keypoints. Int. J. Comput. Vis. 2004, 60, 91–110. [Google Scholar] [CrossRef] [Green Version]
Figure 1. Representation of the target object using structured sparse PCA and deterministic classification between the target object and background image patches.
Figure 1. Representation of the target object using structured sparse PCA and deterministic classification between the target object and background image patches.
Sensors 18 03513 g001
Figure 2. Representation of the target object using structured sparse PCA and deterministic classification between the target object and background image patches.
Figure 2. Representation of the target object using structured sparse PCA and deterministic classification between the target object and background image patches.
Sensors 18 03513 g002
Figure 3. Procedure to find the most similar target object templates using confidence (Equation (9)). (a) Typical explanation to find the target object by weighting the scale factor from positive candidate templates to prevent drift, partial occlusion and scaling problems; (b) real image-based re-weighting procedure to find similar templates from positive image templates.
Figure 3. Procedure to find the most similar target object templates using confidence (Equation (9)). (a) Typical explanation to find the target object by weighting the scale factor from positive candidate templates to prevent drift, partial occlusion and scaling problems; (b) real image-based re-weighting procedure to find similar templates from positive image templates.
Sensors 18 03513 g003
Figure 4. Target models in F at the update time and the detection of the target at a later frame.
Figure 4. Target models in F at the update time and the detection of the target at a later frame.
Sensors 18 03513 g004
Figure 5. Tracking during partial occlusion and drift.
Figure 5. Tracking during partial occlusion and drift.
Sensors 18 03513 g005
Figure 6. Tracking during illumination changes.
Figure 6. Tracking during illumination changes.
Sensors 18 03513 g006
Figure 7. Tracking during background clutter changes.
Figure 7. Tracking during background clutter changes.
Sensors 18 03513 g007
Figure 8. Tracking comparison for the proposed and current state of the art trackers for the Bolt, Lamming, Racecar and Singer image sequences.
Figure 8. Tracking comparison for the proposed and current state of the art trackers for the Bolt, Lamming, Racecar and Singer image sequences.
Sensors 18 03513 g008
Table 1. Notations and symbols.
Table 1. Notations and symbols.
SymbolDescription
I t Frame at time t
X t R 2 × R × R State variable X t = ( x t c , w t s x , h t s y )
Y t { 1 , 1 } Observation variable
x c = ( x c , y c ) R 2 Location vector in the state variable X t
v t g R s Target descriptor vector
v b g R s Background descriptor vector
p Patch image
d Column vectors of D
c Column vectors of C
VFeature descriptor
DFeature dictionary
CFeature coefficient matrices
Φ Support vector machine classifier
σ 4 × 4 Diagonal covariance matrix
N ( X t ; X t 1 , σ ) Multivariate normal distribution
FSet of target descriptors
pProbability function
sDimension of descriptors
rNumber of dictionary vectors
κ Number of background descriptors v b g
kNumber of vectors after updating
tTime variable
a x Real number 1 or 1 related to width size
b y Real number 1 or 1 related to height size
w t s x Width (x-axis) size of patch
h t s y Height (y-axis) size of patch
Approximately equal
Proportional to
· T Transpose operator
Table 2. Average of overlap score of the proposed tracker and several current state of the art trackers ((BC), deformation (DEF), fast motion (FM), in-plane rotation (IPR), illumination variation (IV), low resolution (LR), motion blur (MB), occlusion (OCC), out-of-plane rotation (OPR), out-of view (OV) and scale variation (SV)). The top two methods for each dataset are highlighted in red and blue, respectively. VTD, visual tracking decomposition; MS, mean-shift; MIL, multiple instance learning; SCM, sparse collaborative appearance; Frag, fragment-based; TLD, tracking-learning-detection.
Table 2. Average of overlap score of the proposed tracker and several current state of the art trackers ((BC), deformation (DEF), fast motion (FM), in-plane rotation (IPR), illumination variation (IV), low resolution (LR), motion blur (MB), occlusion (OCC), out-of-plane rotation (OPR), out-of view (OV) and scale variation (SV)). The top two methods for each dataset are highlighted in red and blue, respectively. VTD, visual tracking decomposition; MS, mean-shift; MIL, multiple instance learning; SCM, sparse collaborative appearance; Frag, fragment-based; TLD, tracking-learning-detection.
AllBCDEFFMIPRIVLRMBOCCOPROVSV
Proposed58.5060.1958.7856.7456.2252.0360.8658.9655.0057.1356.5152.33
VTD [30]49.355.146.241.750.253.747.143.552.353.751.548.9
MS [19]35.636.732.840.536.834.628.441.237.437.341.036.0
MIL [22]45.948.645.744.145.747.143.543.747.648.952.744.5
SCM [29]54.461.351.542.851.861.161.745.256.857.056.455.8
Frag [25]44.246.141.844.843.342.642.646.146.646.150.144.2
IVT [24]46.451.640.537.346.451.255.841.349.349.052.347.1
TLD [27]46.848.337.444.648.946.753.351.045.246.050.247.1
Struct [28]57.559.352.455.657.059.059.159.955.957.358.957.8
ASLA [11]53.259.250.542.052.159.659.344.656.056.355.354.0

Share and Cite

MDPI and ACS Style

Yoon, G.-J.; Hwang, H.J.; Yoon, S.M. Visual Object Tracking Using Structured Sparse PCA-Based Appearance Representation and Online Learning. Sensors 2018, 18, 3513. https://doi.org/10.3390/s18103513

AMA Style

Yoon G-J, Hwang HJ, Yoon SM. Visual Object Tracking Using Structured Sparse PCA-Based Appearance Representation and Online Learning. Sensors. 2018; 18(10):3513. https://doi.org/10.3390/s18103513

Chicago/Turabian Style

Yoon, Gang-Joon, Hyeong Jae Hwang, and Sang Min Yoon. 2018. "Visual Object Tracking Using Structured Sparse PCA-Based Appearance Representation and Online Learning" Sensors 18, no. 10: 3513. https://doi.org/10.3390/s18103513

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop