Next Article in Journal
Outage-Based Resource Allocation for DF Two-Way Relay Networks with Energy Harvesting
Previous Article in Journal
THz Spectroscopic Investigation of Wheat-Quality by Using Multi-Source Data Fusion
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Detection and Tracking of Moving Targets for Thermal Infrared Video Sequences

School of Electronic and Information Engineering, Beihang University, Beijing 100191, China
*
Author to whom correspondence should be addressed.
Sensors 2018, 18(11), 3944; https://doi.org/10.3390/s18113944
Submission received: 26 September 2018 / Revised: 4 November 2018 / Accepted: 12 November 2018 / Published: 14 November 2018
(This article belongs to the Section Physical Sensors)

Abstract

:
The joint detection and tracking of multiple targets from raw thermal infrared (TIR) image observations plays a significant role in the video surveillance field, and it has extensive applied foreground and practical value. In this paper, a novel multiple-target track-before-detect (TBD) method, which is based on background subtraction within the framework of labeled random finite sets (RFS) is presented. First, a background subtraction method based on a random selection strategy is exploited to obtain the foreground probability map from a TIR sequence. Second, in the foreground probability map, the probability of each pixel belonging to a target is calculated by non-overlapping multi-target likelihood. Finally, a δ generalized labeled multi-Bernoulli ( δ -GLMB) filter is employed to produce the states of multi-target along with their labels. Unlike other RFS-based filters, the proposed approach describes the target state by a pixel set instead of a single point. To meet the requirement of factual application, some extra procedures, including pixel sampling and update, target merging and splitting, and new birth target initialization, are incorporated into the algorithm. The experimental results show that the proposed method performs better in multi-target detection than six compared methods. Also, the method is effective for the continuous tracking of multi-targets.

1. Introduction

The detection and tracking of moving targets is a challenging vision task that has attracted extensive research. Because of the comparatively lower cost, omnipresence, and 24 × 7 applicability, thermal infrared (TIR) sensors have provided new application areas [1]. Since pedestrians are the major participants in many events of interest, the joint detection and tracking of multi-targets (usually meaning pedestrians, but not exclusively in this paper) becomes one primary task borne by the TIR surveillance system [2]. The main advantages of thermal sensors are their ability to see in complete darkness, their robustness to illumination changes and shadow effects, and their comparatively lower degree of intrusion regarding privacy. Despite many superiorities, the main disadvantages of TIR imaging include low resolution, many dead pixels, lack of color information, low foreground/background contrast, and associated heavy noise [2,3]. As well as the above disadvantages, in top-down surveillance scenes, the target usually occupies fewer pixels and it is difficult to extract the effective appearance features, including textural and contour information. Moreover, the targets in surveillance scenes are highly variable in pose, size, shape, and intensity [2]. Multiple moving-target detection and tracking remain crucial objectives and are identified as the key issues in the TIR surveillance system.
Usually, detection and tracking tasks are separable in computer vision, and most multi-target tracking objectives require a detection operation to produce measurements [4]. Emerging technologies, such as proposal detection method and deep convolution neural network method, usually cannot achieve as admirable detection performance in the TIR surveillance images as the optical images due to the above-mentioned shortcomings. The traditional method still plays a significant role in moving-target detection. In addition, the detection methods in video streams can be divided into three categories: frame difference, optical flow, and background subtraction methods [5]. Because of its low computational cost and high accuracy, the background subtraction method is more popular [6,7]. Unfortunately, the classic background subtraction method has two debilitating drawbacks. First, it achieves detection tasks based on a threshold, which results in information loss [8]. During detection, information loss can significantly degrade tracking performance, especially in obscure feature cases [4,9]. Second, the detection foreground pixels from different targets are not discriminable. To track different targets individually, some post-processes must be executed to differentiate them. Therefore, the joint processing of detection and tracking tasks is of fundamental interest to reduce information loss and simplify the process. Some typical algorithms such as dynamic programming [9,10], Bayesian existence process [11], and multi-modal distributions [12] have shown great success. Among them, the random finite set (RFS) framework approaches that jointly detect and track have attracted significant attention [4,13,14].
The RFS-based methods consist of two procedures in a Bayesian framework: prediction and update. Two methods can be used to implement them, one based on the Gaussian mixtures (GM) model and the other based on the sequential Monte Carlo (SMC) model; the latter is also known as particle implementation. The RFS-based filters can be divided into unlabeled filters and labeled filters. The probability hypothesis density (PHD), cardinalized PHD (CPHD), and multi-target multi-Bernoulli (MeMBer) filters are the typical unlabeled category [15,16,17]. Although these filters have been successfully applied to visual tracking [4,18,19], they provide only unidentified estimates and require additional post-processing to form tracks. The generalized labeled multi-Bernoulli (GLMB) filter and δ -GLMB filter belonging to the label class of RFS-based filters can distinguish and maintain different tracks by adding a label to each target [20,21]. Based on the labeled RFS filter, a δ -GLMB track-before-detect (TBD) approach with a separable likelihood function was introduced in a radar-tracking scenario in [22]. Subsequently, an improved GLMB-TBD algorithm, which can handle non-separable likelihood situations, was proposed in [23] for generic measurement models, although the considerable computational cost limits its application. Currently, the labeled RFS-based methods focus on non-overlapping targets and point target tracking [14,18,22].
To alleviate the above problem in multi-target detection and tracking of TIR surveillance system, a joint detection and tracking approach based on particle implementation, which combines a background subtraction method with a GLMB filter, is proposed. First, according to the ViBe algorithm [24], a random selection strategy background subtraction method without threshold detection is designed to yield a TIR foreground probability map. Second, a multi-target likelihood function is used to calculate the probability of each pixel belonging to a target in the foreground probability map. Finally, the δ -GLMB-TBD filter is applied to produce the state of the multi-target along with their labels. Most RFS-based TBD methods describe a target in the image as a rectangle or a single point [14,18], which is too rough for a target with complex contour and may degrade the detection and tracking performance. In the proposed method, the target is represented by its own pixel set. This means that the δ -GLMB-TBD filter is extended to track irregular areas of a target by benefiting from shape similarity in consecutive frames. To be practical, the algorithm also includes some extra procedures, such as pixel sampling and update, target merging and splitting, and new birth target initialization, which accommodate target deformation and overlapped and dynamic change via gathering, splitting, birth, and death.
The main contribution of this paper is the proposal of a joint multi-target detection and tracking method based on background subtraction and a δ -GLMB-TBD filter for infrared surveillance system, along with its particle implementation. As well as producing a multi-target state estimate, the proposed method can track the multi-targets successfully and individually keep their labels. Based on the proposed method, we also developed the following:
  • More effective multi-target estimates which are from a pixel set instead of a rectangle or a single point;
  • Several procedures to accommodate target deformation and multi-target dynamic processes, such as pixel sampling and update, target merging, splitting, and new target initialization;
  • A random selection strategy background subtraction method which can be used to pre-process the images without threshold segmentation.
This paper is organized as follows. Section 2 describes the proposed algorithm in detail, including the background subtraction method, multi-target likelihood calculation, and the recursion of the δ -GLMB-TBD filter. The results and analysis of the experiments are presented in Section 3. Conclusions are drawn in Section 4.

2. Background-Subtraction-Based δ -GLMB-TBD Filter

In this section, the background-subtraction-based δ -GLMB-TBD filter is introduced in three major parts: background subtraction, multi-target likelihood function calculation, and implementation of the δ -GLMB-TBD filter [22]. A block diagram is presented in Figure 1. Background subtraction transforms the original TIR image to a foreground probability map where each pixel can be interpreted as the probability of the pixel belonging to the foreground. Then the map is used to generate new birth targets and to calculate the multi-target likelihood. Finally, the δ -GLMB-TBD is used to produce multi-target estimates. To accommodate target deformation and dynamic change, some extra procedures, such as splitting and merging, are included.

2.1. Background Subtraction Method for TIR

Figure 2 shows three typical TIR surveillance scenes, in which it is difficult to detect all targets because of low foreground/background contrast, fewer pixels, heavy noise, and lack of textural and contour information. One popular method is background subtraction. However, background subtraction usually produces two segmentations that denote background and foreground by threshold detection, which results in information loss and then leads to inferior tracking performance. To alleviate these problems, the GLMB-TBD filter without the need to detect targets is employed to achieve the joint detection and tracking of multi-target on a background suppression image. Many methods have the potential to subtract the background, such as ViBe [24], KDE [25], GMM [26] and so on. In this paper, we propose a random selection strategy based on ViBe algorithm [24] to subtract the background, which is because the ViBe algorithm has the characters of low complexity, high stability, and excellent background subtraction effect [27]. The new background subtraction method consists of two parts: background model initialization and background model update. The differences between the new approach and the ViBe method are as follows: (1) less missing detection without threshold segmentation; (2) pixel background model update with respect to the probability of being a foreground pixel; (3) morphology operations to eliminate scattered noise and maintain shape. The details are introduced in the following sections.

2.1.1. ViBe Method

Before discussing the proposed method, we will first summarize the standard ViBe method. The background subtraction is regarded as a classification problem in the ViBe method. However, as there is no way to model each background pixel as a probability density function (PDF), then, no estimation or classification result could be given directly from the PDF. Therefore, the ViBe method establishes a background pixel set for each pixel, and each element in this set can be seen as a sample obtained by the true PDF of the background [24].
The key problems of the ViBe method are: (1) how to get the background pixel set, effectively; and (2) how to classify a pixel as a background or foreground according to its given background pixel set. For the first problem, a random selection strategy-based method is employed to update the background pixel set. This can make the samples be more compliant to the pdf of the background without increasing the number of the samples and discarding the earlier samples. For the second problem, the new pixel should compare with its background pixel set. The steps are listed as follows.
Step 1: Calculate the Euclidean distances between the new sample and each sample in its background pixel set;
Step 2: Obtain the number of the Euclidean distances shorter than a given threshold;
Step 3: Compare the number with another given threshold; if the number is greater than the threshold, then the new pixel is classified as a background pixel and vice versa.

2.1.2. Background Model Initialization

Regarding the proposed method, the background model initialization will be discussed first. The proposed method only employs the first frame to initialize the background model. Let y i denote the ith pixel value in the original TIR image I o r i ( 1 i N i m g , where N i m g denotes the number of all pixels in the TIR image), and M i k denotes the background model of pixel i at time k (in the background model initialization, set k = 1 ); all M i k make up the image background model M ( k ) . Each M i k is a collection of N background samples.
M i ( k ) = { m i , 1 , m i , 2 , . . . , m i , N } .
where m i , n ( 1 n N ) is a sample and initialized as follows:
m i , n = y i + v r a n ( v l , v h )
where v r a n ( v l , v h ) denotes a uniform random number between v l and v h . The main parameter in initial model is: N.

2.1.3. Background Model Update

The background model will be updated with each new frame. The update step is the core procedure used to yield accurate results over time. In this step, a conservative update policy is used.
For each pixel i in frame I o r i at time k ( k > 1 ), its equivalent background is the average of its background model M i k given by
y i , e q u = 1 N ( m i , 1 + m i , 2 + . . . + m i , N ) .
We compare the absolute difference between its current value y i and its equivalent y i . e q u with a threshold t h d i f f . The comparison result s i can be obtained by
s i = a b s ( y i - y i , e q u ) / t h d i f f i f a b s ( y i - y i , e q u ) < t h d i f f 1 i f a b s ( y i - y i , e q u ) t h d i f f
where a b s ( · ) denotes the “absolute” operation. s i = 1 means the pixel i is classified as foreground. According to a conservative update policy, a foreground pixel should never be used to update the background. Thus, only when s i < 1 can the current pixel value y i replace one sample in the background model M i k - 1 with the probability of ( 1 - s i ). So t h d i f f plays a significant role in updating the background model. When t h d i f f is too small, s i is sensitive to the pixel change and noise; when t h d i f f is too large, the background model may be polluted by the pixel belonging to moving target. In contrast with the first-in-first-out strategy, the sample substituted by y i is chosen randomly by a uniform probability density function. These operations can extend the time windows covered by the background models. The long lifespan of the background samples significantly aids in the detection of slow-moving multi-targets. The conservative update policy can make a sharp detection of a moving target without introducing the foreground.
Unfortunately, one disadvantage of the conservative update policy is that it can lead to deadlock situations and ghosts. To eliminate these influences, an improved “detection support map” method [28], which counts the number of times a pixel is classified as foreground consecutively, is employed. At frame k, the detection support map D S M i k is given by
D S M i ( k ) = D S M i ( k - 1 ) + 1 s i = 1 0 o t h e r w i s e
We assume the maximum duration of a target remaining stationary is t s t a and that the corresponding frame number is N s t a . If the time of one target remaining stationary reaches t s t a , this target is then classified as a background target. When the D S M i k reaches the threshold N s t a , a random strategy is then used to update the background model. During the update, n r a n samples ( n r a n is a uniform random positive integer between 1 and N) in the background model of pixel i are replaced by new samples. To speed up ghost elimination, when a sample has been updated, one sample from its 8-connected neighborhood pixel background model should be replaced by a new value according to a uniform law. This operation uses the spatial consistency assumption of the background, as in the ViBe method. The assumption is that the background pixel shares a similar distribution as its immediate neighbors. In addition, each new sample is obtained by (2). The detection support map is updated following (6), and the updated background model M i k at time k is obtained.
D S M i ( k ) = D S M i ( k ) - n r a n
Based on the above steps, the whole update background model M ( k ) can be obtained by recursion with the new TIR frame I o r i . Then, (4) can be executed with the updated y i , e q u calculated from M i ( k ) ; all s i ( 1 i N i m g ) constitute the foreground probability map S ( k ) . Figure 3a shows the foreground probability map of Figure 2a. The whole target is separated into several small targets and there is some scattered noise in S ( k ) in Figure 3a. Morphology erosion with small structure can be used to eliminate the scattered noise. Before normalizing the absolute difference by t h d i f f in (4), we conduct erosion and dilation operation on the absolute difference image whose pixel value is a b s y i - y i , e q u . Figure 3b shows the foreground probability map with dilation following erosion. The pseudo-code for the background model update is presented in Algorithm 1. Lines 1–3 determine a pixel to be a foreground pixel ( s i = 1 ) or a background pixel ( s i < 1 ). If the pixel belongs to the background, then we update its background model as lines 4-9. If it belongs to the foreground, besides updating M i ( k - 1 ) lines 11–18 also show how to use the detection support map to eliminate deadlock situations and ghosts.
Algorithm 1: Background model update
Sensors 18 03944 i001

2.2. Multi-Target Likelihood Function Calculation

In the foreground probability map S ( k ) , each s i ( 1 i N i m g ) lies in the interval [0, 1], which can be interpreted as the probability that pixel i should be classified as the foreground. In S ( k ) , each target x illuminates a set of pixels denoted by T ( x ) . Inspired by [4], if a pixel i T ( x ) , its intensity distribution follows the foreground likelihood function g F ( x ) , and if i T ( x ) , its intensity distribution follows the background likelihood function g B ( x ) . These two likelihood functions (or probability density functions of intensity) are of the form:
g F ( x ) = ς F exp ( x x δ F δ F )
g B ( x ) = ς B exp ( - x x δ B δ B )
where δ F and δ B determine the spread rates of the foreground and background intensities respectively, and ς F and ς B are normalizing factors. In general, the ratio of δ F to δ B determines the detection threshold for judging whether a pixel belongs to a target or background. The greater the ratio is, the lower the associated target threshold will be. Because the TBD method allows more suspected targets to be input into the δ -GLMB-TBD filter, δ B should be significantly smaller than δ F . The background intensity remains constant unless it is quite close to the target, whereas the foreground intensity has a significantly more variable and spreading intensity function [4,29]. As the references state, the fluctuation of the ratio does not cause significant changes to the tracking results.
Let s ¯ ( x ) denote the average of all pixels in T ( x ) , i.e.,
s ¯ ( x ) = 1 T ( x ) i T ( x ) s i
where · denotes the cardinality (the number of elements) of a set. We assume that all illumination regions of influences of the multi-target in the TIR image are not overlapped, i.e., x x T ( x ) T ( x ) = . This assumption is reasonable, because an individual cannot recognize the identities and states of overlapped targets. For example, in the TIR image when some pedestrians overlap with each other, an individual cannot determine whether one of the pedestrians has disappeared, whether a pedestrian appears in the surveillance scene, or when or whether the group of pedestrians will separate from each other. In practical application, tracking is a dynamic process. When targets are overlapped, the merging procedure (seen in Section 2.3) will guarantee that all targets are treated as one. After separation, they will be tracked individually by the splitting procedure (seen in Section 2.3). The GLMB-TBD filter will generate the separable likelihood function based on the assumption that the targets do not overlap. Then, if given a target set X with statistically independent pixel values, the likelihood that the set X illuminates the region x X T ( x ) can be expressed as x X g F ( s ¯ ( x ) ) .
Let s ¯ B ( X ) denote the average intensity of the map S ( k ) after filling all target regions with the background pixel value of 0, i.e.,
s ¯ B ( X ) = 1 N i m g ( i = 1 N i m g s i - x X i T ( x ) s i ) .
Substituting s ¯ B ( X ) into (8),
g B ( s ¯ B ( X ) ) = ς B exp - i = 1 N i m g s i - x X i T ( x ) s i δ B N i m g = ς B exp - i = 1 N i m g s i δ B N i m g x X exp i T ( x ) s i δ B N i m g = ς B exp - i = 1 N i m g s i δ B N i m g x X exp T ( x ) · s ¯ ( x ) δ B N i m g .
Then, given the target set X, the multi-target likelihood of S ( k ) is the product of the foreground and background, i.e.,
g ( S ( k ) | X ) = g B ( s ¯ B ( X ) ) x X g F ( s ¯ ( x ) ) = ς B exp - i = 1 N i m g s i δ B N i m g i n d e p e n d e n t o f X x X exp T ( x ) · s ¯ ( x ) δ B N i m g g F ( s ¯ ( x ) ) d e p e n d e n t o n x .
(12) shows that this likelihood is separable.

2.3. δ -GLMB-TBD Filter

After obtaining the multi-target likelihood for the foreground probability map, we describe how to detect and track multi-targets using the improved δ -GLMB-TBD filter with a separable likelihood function [30]. The main contribution in this part is that the improved filter can produce the appearance of the target in contrast to the center position or rectangle estimates obtained by the standard δ -GLMB-TBD filter. Based on this, we also develop the following: (1) merging and splitting procedures are employed to handle situations where multi-targets merge into one group and one group splits into several multi-targets; (2) pixel sampling and updating are used to accommodate target deformation; (3) birth target initialization procedure is to open up the applications of the filter. We discuss each of these improvements in the following subsections.

2.3.1. Basic Theory

In this subsection, the standard δ -GLMB-TBD filter, consisting of two steps, prediction and update, with a separable likelihood function, is briefly discussed [23]. Before introducing the recursion of the standard δ -GLMB-TBD filter, some notations are shown for convenience.
(1) Notation
For the remainder of the paper, let lowercase letters (e.g., x) denote single-target state and uppercase letters (e.g., X) denote multi-target states. The labeled target states are indicated by boldface letters (e.g., x , X ). Space is represented by a letter with a tilde (e.g., X ˜ denotes the state space, L ˜ denotes the discrete label space). A labeled target can be written as x = ( x , l ) , where l L ˜ and l = ( k , i ) , k means the target birth time, and i is a unique index to distinguish targets born at the same time. A labeled multi-Bernoulli (LMB) RFS with state space X ˜ and label space L ˜ can be written as v = { r ( l ) , p ( l ) } l L ˜ , where r ( l ) and p ( l ) mean the existence probability and the probability density of a target with label l. α , β = α ( x ) β ( x ) d x denotes the standard inner product, and h X = x X h ( x ) denotes the multi-target exponential, where h = 1 by convention and α ( x ) , β ( x ) , and h ( x ) are real-valued functions. The generalized Kronecker delta function and inclusion function are defined as
δ Y ( X ) = 1 i f X = Y 0 o t h e r w i s e ,
1 Y ( X ) = 1 i f X Y 0 o t h e r w i s e .
(2) Standard δ -GLMB-TBD Filter
From [23], the multi-target posterior at time k has the following δ -GLMB form:
v k ( X ) = Δ ( X ) I F ( L ˜ 0 : k ) δ I ( L ( X ) ) w k ( I ) [ p k ( I ) ] X
where X is the current multi-target state; Δ ( X ) = δ X ( L ( X ) ) denotes the distinct label indicator, which means that the cardinalities of the set of labels and the set of state vectors are identical; L ˜ 0 : k denotes the label space of targets born between time 0 and time k, where the subscript 0:k means time interval [ 0 , k ] ; F ( · ) denotes collections of all finite subsets of a given space; L is a projection from space X ˜ × L ˜ to L ˜ and hence L ( X ) = { L ( x ) : x X } is the set of labels of X ; and w k ( I ) denotes the joint existence probability of the label set I, while the multi-target exponential [ p k ( I ) ] X denotes the joint probability density of X , conditional on their corresponding label set I.
The new birth model covering labeled Poisson, labeled identically and independently distributed cluster and labeled multi-Bernoulli filter can be given by [20]
f B = Δ ( Y ) w B ( L ( Y ) ) [ p B ] Y .
where Y is the state of new birth targets, and w B and [ p B ] Y are the joint existence probability and probability density. This model can also be written as the LMB birth model [23,31] as follows:
w B ( L ) = i L ˜ k ( 1 - r B ( i ) ) l L 1 L ˜ k ( l ) r B ( l ) 1 - r B ( l )
p B ( x , l ) = p B ( l ) ( x )
where r B ( l ) and p B ( l ) ( x ) mean the existence probability and the probability density of a birth target with label l.
Proposition 1.
If the multi-target state posterior with δ-GLMB form is given as (15) at time k, with the new birth model (16), the multi-target prediction density also has a δ-GLMB form [23]:
v k + 1 | k ( X ) = Δ ( X ) I F ( L ˜ 0 : k + 1 ) δ I ( L ( X ) ) w k + 1 | k ( I ) [ p k + 1 | k ( I ) ] X
where
w k + 1 | k ( I ) = w S ( I ) ( I L ˜ 0 : k ) w B ( I L ˜ k + 1 )
w S ( I ) ( L ) = [ η S ( I ) ] L J L ˜ 0 : k 1 J ( L ) [ 1 - η S ( I ) ] J - L w k ( I ) ( J )
p k + 1 | k ( I ) ( x , l ) = 1 L ˜ 0 : k ( l ) P S ( I ) ( x , l ) + ( 1 - 1 L ˜ 0 : k ( l ) ) p B ( x , l )
P S ( I ) ( x , l ) = p S ( · , l ) f k + 1 | k ( x | · , l ) , p k ( I ) ( · , l ) η S ( I ) ( l )
η S ( I ) ( l ) = p S ( · , l ) , p k ( I ) ( · , l )
f k + 1 | k ( · ) denotes a single-target transition function. p S ( x , l ) denotes the survival probability of target x. L ˜ k + 1 is the label space of a target born at time k + 1 .
Proposition 2.
If the multi-target prediction density has the form of δ-GLMB as (19), then, with the measurement set S and separable likelihood function γ S ( x ) , the multi-target posterior density also has the same form [23]:
v k + 1 ( X | S ) = Δ ( X ) I F ( L ˜ 0 : k + 1 ) δ I ( L ( X ) ) w k + 1 ( I ) ( S ) [ p k + 1 ( I ) ( · | S ) ] X
where
w k + 1 ( I ) ( S ) w k + 1 | k ( I ) [ η S ] I
p k + 1 ( I ) ( x , l | S ) = p k + 1 | k ( I ) ( x , l ) γ S ( x , l ) p k + 1 | k ( I ) ( x , l ) γ S ( x , l ) η S ( l ) η S ( l )
η S ( l ) = p k + 1 | k ( I ) ( · , l ) , γ S ( · , l ) .
Proposition 3.
If the multi-target state posterior with the δ-GLMB form is given as (15) at time k, the cardinality distribution ρ k ( n ) can be given by [20]
ρ k ( n ) = I F ( L ˜ 0 : k ) δ n ( I ) w k ( I ) .
The cardinality estimates can be obtained by
n ^ = arg max n ρ k ( n )
The multi-target state estimate is the mean estimate of the multi-target state conditioned on the estimated cardinality n ^ as in [20].

2.3.2. Recursion

Particle implementation is based on the standard δ -GLMB filter implementation [20,21]. Each target density p k ( l ) with label l in p k ( I ) is modeled as a set of weighted samples { ( Ω k , n ( l ) , x k , n ) } n = 1 J k ( l ) , where J k ( l ) is the number of particles, and x n denotes the state x n and label l. Besides the prediction and update steps, we add four other parts: pruning, splitting, merging, and new birth target initialization to form the whole solution. The details are as follows.
(1) Prediction
The pixels illuminated by the target can describe the target more effectively than a single point or a rectangle. Let us assume the single-target state x have two fields: a geometric center location x . c e n and cover area x . c o v , which can be obtained from T ( x ) . For convenience, let x . c e n = x . c e n and x . c o v = x . c o v in this paper. This new approach representing a single-target benefits from the similar target shape in consecutive frames. To accommodate the small amount of deformation, pixel sampling is executed in this prediction step. Each pixel in x . c o v has the probability of p p i x to be selected to survive, the value of which depends on the deformation degree. In general, when the deformation is small, p p i x is close to 1. It is obvious that sampling may reduce the covered area in the final estimates. Fortunately, this can be alleviated in the update step.
If at time k, the multi-target state posterior is given by (15), when the LMB new birth model { ( r B , k + 1 ( l ) , p B , k + 1 ( l ) ) } l L ˜ k + 1 is known (in practice for the unknown birth model, the third part in this subsection describes how to initialize the new birth model), where p B , k + 1 ( l ) is { ( Ω B , k + 1 , n ( l ) , x B , k + 1 , n ) } n = 1 B ( l ) and B ( l ) is the number of particles, according to proposition 1, we obtain
η S ( I ) ( l ) = n = 1 J k ( l ) Ω k , n ( l ) p S ( x k , n )
and p k + 1 | k ( I ) ( x , l ) can be represented as
{ ( 1 L ˜ 0 : k ( l ) Ω ˜ S , k + 1 | k , n ( l ) , x S , k + 1 | k , n ) } i = n J k ( l ) { ( 1 L ˜ 0 : k ( l ) Ω B , k + 1 , n ( l ) , x B , k + 1 , n ) } i = n B ( l ) ,
where x S , k + 1 | k , n . c e n q ( · | x k , n . c e n ) ( n = 1 , 2 , . . . , J k ( l ) ), x S , k + 1 | k , n . c o v is generated by a random pixel sample as described above; Ω S , k + 1 | k , n ( l ) = Ω k , n ( l ) f k + 1 | k ( x S , k + 1 | k , n | x k , n , l ) p S ( x k , n , l ) q ( x S , k + 1 | k , n . c e n | x k , n . c e n ) ; Ω ˜ S , k + 1 | k , n ( l ) = Ω S , k + 1 | k , n ( l ) Ω S , k + 1 | k , n ( l ) n = 1 J k ( l ) Ω S , k + 1 | k , n ( l ) / n = 1 J k ( l ) Ω S , k + 1 | k , n ( l ) ; and q ( · | x k , n . c e n ) is a proposal density. The procedure for calculating w k + 1 | k ( I ) in (20) is bothersome and complex; however, reference [21] offers a method based on a K-shortest paths algorithm to carry it out. To understand the procedure more clearly, reference [21] is strongly recommended.
(2) Update
With the assumption that targets are not overlapped, we obtain a separable likelihood function shown as (12). After adding a distinct label to each target, (12) can be written as
g ( S ( k ) | X ) x X γ S ( x )
where γ S ( x ) = g F ( s ¯ ( x ) ) .
According to proposition 2, if each single-target density p k + 1 | k ( l ) is modeled by a particle set { ( Ω k + 1 | k , n ( l ) , x k + 1 | k , n ) } n = 1 J k + 1 | k ( l ) . Then,
η S ( l ) = n = 1 J k + 1 | k ( l ) w k + 1 | k , n ( l ) γ S ( x k + 1 | k , n )
and p k + 1 ( I ) ( x , l | S ) can be represented as
{ ( Ω k + 1 , n ( l ) , x k + 1 , n ) } n = 1 J k + 1 ( l ) ,
where w k + 1 , n ( l ) = w k + 1 | k , n ( l ) γ S ( x k + 1 | k , n ) / n = 1 J k + 1 | k ( l ) w k + 1 | k , n ( l ) γ S ( x k + 1 | k , n ) ; x k + 1 , n = x k + 1 | k , n ; and J k + 1 ( l ) = J k + 1 | k ( l ) . Substituting (34) into (26), we obtain w k + 1 ( I ) ( S ) .
Analogous to the standard particle filter, resampling each target density { ( Ω k + 1 , n ( l ) , x k + 1 , n ) } n = 1 J k + 1 ( l ) must be executed to reduce the degeneracy. For simplicity, in this paper, multinomial resampling is used for numerical studies that would otherwise be carried out with other multi-Bernoulli filters [17].
To eliminate the influence of pixel sampling in the prediction step, a pixel set update procedure is used to correct the shape of the target. Taking target x as an example, the procedure is described as follows. First, for all J k + 1 ( l ) particles representing target x, count the pixel i ( i j = 1 J k + 1 ( l ) T ( x k + 1 , j ) ) occupied times O T i J k + 1 ( l ) . The occupied times are initialized to 0, i.e., O T i 0 = 0 , and for the jth ( 1 j J k + 1 ( l ) ) particle state x k + 1 , j , O T i j is given by
O T i j = O T i j - 1 + 1 i f i T ( x k + 1 , j ) O T i j - 1 o t h e r w i s e .
Second, compare the O T i J k + 1 ( l ) with a threshold p p i x J k + 1 ( l ) p p i x J k + 1 ( l ) 2 2 . When O T i J k + 1 ( l ) reaches the threshold, the pixel i is classified as an occupied pixel. Third, if the occupied pixel i is greater than an extremely low value in the foreground probability map, then it is considered to be a pixel in the updated pixel set for target x. The updated pixel set is the estimated cover area of target x. The pseudo-code for the pixel set update procedure is presented in Algorithm 2. Algorithm 2 shows us that for each pixel belonging to the covered areas occupied by all particles representing x, the occupied time is calculated (line 3) and then it will be used to determine the true cover area of the target (lines 5–9). The filter will propagate a pixel set as the target state in the recursion and will directly produce the pixel set as its output. This operation can facilitate target recognition and extraction and subsequent processing in computer vision applications.
Algorithm 2: Pixel set update
Sensors 18 03944 i002
(3) Pruning, Splitting, Merging and Birth Target Initialization
To reduce computational complexity, besides the truncation [20,21], the pruning procedure is also needed. The multi-target set X should be discarded if its weight w k + 1 ( I ) is less than a threshold t h p r . In practical application, the dynamic processes of multi-targets are complex. To accommodate these processes, we propose splitting and merging, which are described as follows. For each target x , clustering is executed to check whether the target x should be split into several small targets. If yes, the small target with the most pixels inherits the identity (label) of target x , while the others are labeled as new birth targets. For two targets in X with substantial overlap, if the overlap ratio of the pixel intersection area to the smaller target area is higher than a threshold t h m e , these two targets should be merged; the merged label is the same as the earlier born target.
Birth target initialization has two pre-processing steps. Step 1: clustering the foreground probability map to obtain current cluster target x c , i ( 0 i n , n denotes the number of all current cluster targets). Step 2: for each cluster target x c , i , calculate the overlap ratio between the x c , i and all existing targets. If the maximum overlap ratio is higher than a threshold t h b i r , then remove this x c , i . Subsequently, we use the remaining x c , i to initiate the new birth target LMB modeled as v B , k + 1 = { ( r B , k + 1 ( l ) , p B , k + 1 ( l ) ) } l L ˜ k + 1 . For simplicity, the r B , k + 1 ( l ) is initialized as a constant and the target density p B , k + 1 ( l ) is modeled as { ( Ω B , k + 1 , n ( l ) , x B , k + 1 , n ) } n = 1 B ( l ) , where all Ω B , k + 1 , n ( l ) have equal weight; and x B , k + 1 , n . c e n N ( u B ( l ) , Q B ( l ) ) , where N ( · ) denotes the Gaussian distribution, u B ( l ) is the geometric center location of one remaining x c , i , and Q B ( l ) is the variance. x B , k + 1 , n . cov can be obtained from T ( x c , i ) using random sampling by probability p p i x .
These procedures are designed to broaden the applications of the filter. Without splitting and merging, this filter would produce a cardinality estimation error. Missing the new target birth process would invalidate the δ -GLMB-TBD filter. Using proposition 3 can produce a multi-target estimate.

3. Experimental Results

In this section, the optimal parameters of the background-subtraction-based δ -GLMB-TBD filter are determined. Based on the optimal parameters, detection performance is compared to several state-of-the-art techniques. Then, tracking performance is tested. Based on experimental results, the advantages and disadvantages of the proposed algorithm are discussed. In these experiments, the kinematic transition is modeled as the random walk model; the survival probability of a target is 0.98; the maximum number of frames in which the target remains stationary N s t a is 60; the morphology erosion and dilation operators are executed once with the same flat disk-shaped structuring element whose radius is 1. The particle number for each target is 200; the pruning threshold t h p r is 0.01; the DBSCAN method is chosen for clustering [32]; the threshold t h b i r in birth targets initialization is 0.5; the merging threshold t h m e is 0.7; the existence probability r B , k + 1 ( l ) is 0.3; and the variance Q B ( l ) is [1,0;0,1]. Theoretically, the same detection result and data association can achieve the same tracking performance. Therefore, the detection performance is the main evaluation criteria for the joint detection and tracking method. This novel TBD filter emphasizes detection by exploiting the trajectory information. The measure metrics of recall, precision, and F-measure (FM) are used to evaluate performance, and they are defined as:
r e c a l l = t p t p + f n , p r e c i s i o n = t p t p + f p , F M = 2 · p r e c i s i o n · r e c a l l p r e c i s i o n + r e c a l l
where t p denotes the number of true positives, and f p and f n represent the number of false positives and false negatives.

3.1. Determination of Parameters

From previous discussions, the parameters to be determined are as follows: (1) the threshold t h d i f f in (4), (2) the number of samples N, (3) δ F and δ B , and (4) pixel sampling rate p p i x . The tested sequence is OTCBVS Dataset 03 2a [33]. The fluctuation of the δ F to δ B ratio does not cause obvious changes in the tracking result, where δ F = 0 . 2 and δ B = 0 . 05 are suitable for our experiments.
The proposed method can be divided into two parts: detection and tracking, which are connected by the foreground probability map S ( k ) . The two parts can be considered as conditional independence (CI) based on S ( k ) . Therefore, the parameters also can be divided into two groups: (1) t h d i f f and N (before obtaining S ( k ) ); (2) p p i x (after obtaining S ( k ) ). Let us use the “trial-and-error” method to determine the two groups of parameters respectively as follows.
To choose the optimal value for t h d i f f and N, the performance metrics are calculated for t h d i f f as: 20, 40, 60, 80, and 100 while N as: 10, 20, 30, 40, and 50. The other parameters are fixed at δ F = 0.2, δ B = 0.05 and p p i x = 0.8. The results are shown in Table 1, Table 2 and Table 3. From Table 1, we can see that the recall decreases with increasing t h d i f f . When t h d i f f is set 20, the recall can achieve the highest value. The number of samples N has less influence on recall; in general, with the increase of N, the recall tends to become higher. From Table 2, the result of precision is opposite to the recall metric. Table 3 shows that the balance metric F M has the highest value at t h d i f f = 60. In this TBD method, recall plays a more important role than precision. The range of t h d i f f can be set between 40 and 60, N is between 20 and 40.
Now, let us determine the pixel sampling rate p p i x . When we set t h d i f f to 60, δ F to 0.2, δ B to 0.05, and N to 20, Table 4 shows the experimental results. From Table 4, the evaluation metrics do not change significantly, because the p p i x mainly affects the detection and tracking of dim targets that occupy few pixels and may be deformed in consecutive frames. A p p i x closer to 1 is of minimal help in detecting and tracking deformation targets, and a smaller p p i x may cause a dim target to be undetected. In the experiment, the p p i x is set to 0.8.

3.2. Comparison with Other Techniques

To the best of our knowledge, there is no common dataset for TIR multiple moving-target detection and tracking in the surveillance scene. We collected 15 sequences containing about 14,753 images in 7 different locations at different times from OTCBVS [34] (IEEE OTCBVS WS Series Bench and Roland Miezianko, Terravic Reasearch Infrared Database) and PTB-TIR [35]. The moving targets are mainly pedestrians, but also include pets, cars, and trucks. In this subsection, the performance of the proposed method is compared with 6 methods: ViBe [24], pixel-based adaptive segmenter (PBAS) [36], and kernel density estimator (KDE) [25], SLIC-based method [37], orthogonal rank-one matrix pursuit method (OR1MP) [38], and robust PCA via Gradient Descent (RPCA-GD) [39]. The ViBe, PBAS, and KDE are classic state-of-the-art methods. The SLIC, OR1MP, and RPCA-GD are 3 new methods. Among them, SLIC is a proposal method designed for infrared target detection; OR1MP and RPCA-GD can be recursive and unable to give instantaneous detection because of the non-causality. The programs of the ViBe, KDE, OR1MP, and RPCA-GD methods are provided by their authors, and the parameters used in the experiments are suggested by the authors. The PBAS method is available in the BGSlibrary [40]. SLIC is implemented by us using MATLAB; we define t h d = m e a n s + 4 s t d s as the detection threshold, where means and stds are the mean and the standard deviation of all saliency scores, respectively.
First, we use four typical sequences to validate the effectiveness of the proposed method. They cover many typical situations such as targets crossing, entering/leaving, gathering/separation, occlusion, stopping /restarting, and irregular motion. The sequences are divided into two categories: sparse target scenarios and dense target scenarios. The results of the four sequences are well represented for examining the performance of the proposed method. The parameters used in our model are t h d i f f = 60, N = 20 , δ F = 0 . 2 , δ B = 0 . 05 , and p p i x = 0 . 8 . For the proposed method, the output results are the estimates of the filter, while for other methods, the results are the foreground detections.
Figure 4, Figure 5, Figure 6 and Figure 7 shows examples of moving-target detection for four typical frames chosen from four different sequences. Table 5 presents the average evaluation metrics of the seven methods. For all four sequences, the proposed method can detect all moving targets, despite some false alarms. Compared with our method, ViBe detects more false alarms and PBAS produces missing detections. The KDE produces obvious worse detection results than the proposed method. Because the proposal-based method only uses the features of the target to detect and pay no attention to the information in inter frame, SLIC can only detect the highlight areas in the images. In general, other algorithms that do not use information between consecutive frames, such as the SLIC method, will obtain similar results. OR1MP and RPCA-GD can also obtain good detection performance on Seq. 2; this is because their non-causality allows them to use the follow-up frames to build the current time model to eliminate the ghost. The non-causality will limit their application. However, for Seq.1, OR1MP and RPCA-GD detection results occupy more pixels than the ground truth; for Seq. 3 and Seq. 4 they produce obvious missing detection. These results are validated by the metric scores in Table 5. The good performance of the proposed method benefits from the detection support map, the neighbor pixel update processing, and the use of trajectory information. Figure 4, Figure 5, Figure 6 and Figure 7 and Table 5 indicate that the proposed method outperforms the other six methods obviously.
In addition, Table 6 shows average evaluation metrics and runtime per frame for all 15 sequences. From Table 6, we can see that the proposed method can get the highest recall, precision, and FM metrics even compared with the non-causal method. Also, Table 6 validates that the proposed method has the best detection performance among the 7 methods.

3.3. Discussion of Tracking Performance

The background-subtraction-based δ -GLMB-TBD method can also produce target trajectories without additional post-processing, and accommodate multi-target dynamic process and measurement process, although its main task is target detection. In this section, the optimal sub-pattern assignment (OSPA) metric [41] interpreted as per-target tracking error containing cardinality error and state estimation error with parameters p = 1 and c = 50 pixel will be employed as the main performance metric.
In this section, the tracking performance of the proposed method is compared with the standard Kalman filter whose detections are provided by ViBe and PBAS methods. The models and parameters used in Kalman filter are the same as the proposed method, but in Kalman filters, the target state is represented by its centroid. The cardinality estimates and OSPA curves tested on Seq. 1 and Seq. 4 are shown in Figure 8 and Figure 9. From Figure 8; we can see that without overlapped targets, our method and PBAS + Kalman method can almost produce the correct cardinality; and with complex multi-target motion and occlusion, the 3 filters produce obvious cardinality estimation error, but they still can reflect the trend of cardinality change. From Figure 9, we can see, in both simple and complex scenarios, the proposed method can obtain the lowest OSPA value, this means the proposed method has the best multi-target tracking performance in the 3 methods. This is because the proposed method can yield accurate appearance and centroid estimation. The average OSPA values of all 15 sequences is shown in Table 7. Also, Table 7 indicates the proposed method can obtain the best multi-target tracking performance.
Figure 10 shows two typical frames with estimation error. The proposed method outputs targets that are filled with different colors. From Figure 10, we can see the “ghost” (seen in Figure 10b), similarity between the targets and background (red rectangles with number 1, 2, 6, 7 in Figure 10b) can cause the over estimation; occlusions between targets or between target and background (red rectangles with number 6, 9 in Figure 10b) is the main reason to produce low estimation.
As described above, this new filter can add a label to the target to maintain the track. Figure 11 shows the estimated trajectories of moving targets whose duration exceeds 30 frames. The different colors in Figure 11 denote different trajectories. According to Figure 11, the filter can track the targets successfully along with their labels without overlap or occlusion. When a target is separated from the crowd, this new filter can continue tracking it as a new target. The estimates from the labeled filter can facilitate tracking, recognition, and other subsequent processing in computer vision. In future, the trace association could be used to merge the tracks before and after occlusion to form a long-time trajectory.

4. Conclusions

A novel method for moving-target detection and tracking directly from the TIR sequence in surveillance scenes was proposed based on background-subtraction and the δ -GLMB-TBD filter. First, a background subtraction method using a random selection strategy was used to produce the foreground probability map. Separable non-overlapped multi-target likelihood was exploited to obtain the probability of the pixels belonging to the foreground. Then, the δ -GLMB-TBD filter was used to provide estimates. Unlike other RFS-based filters, the proposed method used the pixel set, which was the target projection in the image, to describe the target instead of a rectangle or a single point. This means the δ -GLMB-TBD filter directly produced a continuous trajectory as well as accurate multi-target shape estimates. In implementation, several procedures including pixel sampling and update, target merging and splitting, and new birth target initialization were combined in the method to accommodate target deformation and multi-target dynamic change: gathering, splitting, birth, and death. After describing the method, the optimal parameters were determined by experiments. Then, the performance of the novel method was compared with six existing methods. According to the experimental results, the proposed TBD method obtained the highest FM scores, meaning that it outperformed the other six methods in the detection of moving targets. The experiments also show that the proposed method can achieve better tracking performance than the Kalman filters with different detections. In future, the proposed method could be extended to the detection and tracking of moving targets without non-overlap assumptions.
References
References

Author Contributions

Methodology, C.L. and W.W.; Validation, C.L. and W.W.; Formal Analysis, C.L.; Writing—Original Draft Preparation, C.L. and W.W.; Writing—Review & Editing, W.W. and C.L.; Project Administration, W.W.; Funding Acquisition, W.W.

Funding

This research was funded by the National Natural Science Foundation of China [61771028].

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Gade, R.; Moeslund, T.B. Thermal cameras and applications: A survey. Mach. Vis. Appl. 2014, 25, 245–262. [Google Scholar] [CrossRef]
  2. Tan, Y.; Guo, Y.; Gao, C.; Tan, Y.; Guo, Y.; Gao, C. Background subtraction based level sets for human segmentation in thermal infrared surveillance systems. Infrared Phys. Technol. 2013, 61, 230–240. [Google Scholar] [CrossRef]
  3. Berg, A.; Ahlberg, J.; Felsberg, M. Channel Coded Distribution Field Tracking for Thermal Infrared Imagery. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Las Vegas, NV, USA, 26 June–1 July 2016; pp. 1248–1256. [Google Scholar]
  4. Hoseinnezhad, R.; Vo, B.N.; Vo, B.T. Visual Tracking in Background Subtracted Image Sequences via Multi-Bernoulli Filtering. IEEE Trans. Signal Process. 2013, 61, 392–397. [Google Scholar] [CrossRef]
  5. Shaikh, S.H.; Saeed, K.; Chaki, N. Moving Object Detection Using Background Subtraction; Springer International Publishing: Cham, Switzerland, 2014; pp. 15–23. [Google Scholar]
  6. St-Charles, P.L.; Bilodeau, G.A.; Bergevin, R. Universal Background Subtraction Using Word Consensus Models. IEEE Trans. Image Process. 2016, 25, 4768–4781. [Google Scholar] [CrossRef]
  7. Gemignani, G.; Rozza, A. A Robust Approach for the Background Subtraction Based on Multi-Layered Self-Organizing Maps. IEEE Trans. Image Process. 2016, 25, 5239–5251. [Google Scholar] [CrossRef] [PubMed]
  8. Hurney, P.; Waldron, P.; Morgan, F.; Jones, E.; Glavin, M. Review of pedestrian detection techniques in automotive far-infrared video. Iet Intell. Transp. Syst. 2015, 9, 824–832. [Google Scholar] [CrossRef]
  9. Yi, W.; Morelande, M.R.; Kong, L.; Yang, J. An Efficient Multi-Frame Track-Before-Detect Algorithm for Multi-Target Tracking. J. Sel. Top. Signal Process. 2013, 7, 421–434. [Google Scholar] [CrossRef]
  10. Jiang, H.; Yi, W.; Kirubarajan, T.; Kong, L.; Yang, X. Track-Before-Detect Strategies for Radar Detection in G0-distributed Clutter. IEEE Trans. Aerosp. Electron. Syst. 2017, 53, 2516–2533. [Google Scholar]
  11. Vermaak, J.; Maskell, S.; Briers, M.; Perez, P. Bayesian visual tracking with existence process. In Proceedings of the IEEE International Conference on Image Processing, San Diego, CA, USA, 12–15 October 2008; pp. I-721–I-724. [Google Scholar]
  12. Czyz, J.; Ristic, B.; Macq, B. A particle filter for joint detection and tracking of color objects. Image Vis. Comput. 2007, 25, 1271–1281. [Google Scholar] [CrossRef]
  13. Punithakumar, K.; Kirubarajan, T. A sequential Monte Carlo probability hypothesis density algorithm for multitarget track-before-detect. In Proceedings of the SPIE—The International Society for Optical Engineering, Boston, MA, USA, 24–25 October 2005; pp. 59131S–59131S-8. [Google Scholar]
  14. Li, M.; Li, J.; Zhou, Y. Labeled RFS-Based Track-Before-Detect for Multiple Maneuvering Targets in the Infrared Focal Plane Array. Sensors 2015, 15, 30839–30855. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  15. Mahler, R.P.S. Multitarget Bayes filtering via first-order multitarget moments. IEEE Trans. Aerosp. Electron. Syst. 2004, 39, 1152–1178. [Google Scholar] [CrossRef]
  16. Mahler, R. PHD filters of higher order in target number. IEEE Trans. Aerosp. Electron. Syst. 2008, 43, 1523–1543. [Google Scholar] [CrossRef]
  17. Vo, B.T.; Vo, B.N.; Cantoni, A. The Cardinality Balanced Multi-Target Multi-Bernoulli Filter and Its Implementations. IEEE Trans. Signal Process. 2009, 2, 409–423. [Google Scholar]
  18. Hoseinnezhad, R.; Vo, B.N.; Vo, B.T.; Suter, D. Visual tracking of numerous targets via multi-Bernoulli filtering of image data. Pattern Recognit. 2012, 45, 3625–3635. [Google Scholar] [CrossRef]
  19. Zhou, X.; Li, Y.F.; He, B. Entropy distribution and coverage rate-based birth intensity estimation in GM-PHD filter for multi-target visual tracking. Signal Process. 2014, 94, 650–660. [Google Scholar] [CrossRef]
  20. Vo, B.T.; Vo, B.N. Labeled Random Finite Sets and Multi-Object Conjugate Priors. IEEE Trans. Signal Process. 2013, 61, 3460–3475. [Google Scholar] [CrossRef]
  21. Vo, B.N.; Vo, B.T.; Phung, D. Labeled Random Finite Sets and the Bayes Multi-Target Tracking Filter. IEEE Trans. Signal Process. 2014, 62, 6554–6567. [Google Scholar] [CrossRef] [Green Version]
  22. Papi, F.; Vo, B.T.; Bocquel, M.; Vo, B.N. Multi-target Track-Before-Detect using labeled random finite set. In Proceedings of the 2013 International Conference on Control, Automation and Information Sciences (ICCAIS), Nha Trang, Vietnam, 25–28 November 2013; pp. 116–121. [Google Scholar]
  23. Papi, F.; Vo, B.N.; Vo, B.T.; Fantacci, C.; Beard, M. Generalized Labeled Multi-Bernoulli Approximation of Multi-Object Densities. IEEE Trans. Signal Process. 2015, 63, 5487–5497. [Google Scholar] [CrossRef]
  24. Barnich, O.; Droogenbroeck, M.V. ViBe: A Universal Background Subtraction Algorithm for Video Sequences. IEEE Trans. Image Process. 2011, 20, 1709. [Google Scholar] [CrossRef] [PubMed]
  25. Elgammal, A.; Duraiswami, R.; Harwood, D.; Davis, L.S. Background and foreground modeling using nonparametric kernel density estimation for visual surveillance. Proc. IEEE 2002, 90, 1151–1163. [Google Scholar] [CrossRef] [Green Version]
  26. Zivkovic, Z. Improved adaptive Gaussian mixture model for background subtraction. In Proceedings of the 17th International Conference on Pattern Recogntion, Cambridge, UK, 23–26 August 2004; Volume 2, pp. 28–31. [Google Scholar]
  27. Xu, Y.; Dong, J.; Zhang, B.; Xu, D. Background modeling methods in video analysis: A review and comparative evaluation. CAAI Trans. Intell. Technol. 2016, 1, 43–60. [Google Scholar] [CrossRef]
  28. Haritaoglu, I.; Harwood, D.; David, L.S. W4: Real-Time Surveillance of People and Their Activities; IEEE Computer Society: Washington, DC, USA, 2000; pp. 809–830. [Google Scholar]
  29. Punchihewa, Y.; Papi, F.; Hoseinnezhad, R. Multiple target tracking in video data using labeled random finite set. In Proceedings of the International Conference on Control, Automation and Information Sciences, Gwangju, South Korea, 2–5 December 2014; pp. 13–18. [Google Scholar]
  30. Reuter, S.; Vo, B.T.; Vo, B.N.; Dietmayer, K. The Labeled Multi-Bernoulli Filter. IEEE Trans. Signal Process. 2014, 62, 3246–3260. [Google Scholar]
  31. Papi, F.; Gostar, A.K. Bayesian Track-Before-Detect for closely spaced targets. In Proceedings of the IEEE 2010 International Conference on Signal Processing Conference, Beijing, China, 24–28 October 2010; pp. 1979–1983. [Google Scholar]
  32. Ester, M.; Kriegel, H.P.; Xu, X. A density-based algorithm for discovering clusters a density-based algorithm for discovering clusters in large spatial databases with noise. In Proceedings of the International Conference on Knowledge Discovery and Data Mining, Portland, OR, USA, 2–4 August 1996; pp. 226–231. [Google Scholar]
  33. Davis, J.W.; Sharma, V. Background-subtraction using contour-based fusion of thermal and visible imagery. Comput. Vis. Image Underst. 2007, 106, 162–182. [Google Scholar] [CrossRef]
  34. Wu, Z.; Fuller, N.; Theriault, D.; Betke, M. A thermal infrared video benchmark for visual analysis. In Proceedings of the 10th IEEE Workshop on Perception Beyond the Visible Spectrum, Columbus, OH, USA, 23–28 June 2014. [Google Scholar]
  35. Liu, Q.; He, Z. PTB-TIR: A Thermal Infrared Pedestrian Tracking Benchmark. arXiv, 2018; arXiv:1801.05944. [Google Scholar]
  36. Hofmann, M.; Tiefenbacher, P.; Rigoll, G. Background segmentation with feedback: The Pixel-Based Adaptive Segmenter. In Proceedings of the Computer Vision and Pattern Recognition Workshops, Providence, RI, USA, 16–21 June 2012; pp. 38–43. [Google Scholar]
  37. Sun, N.; Jiang, F.; Yan, H.; Liu, J.; Han, G. Proposal generation method for object detection in infrared image. Infrared Phys. Technol. 2017, 81, 117–127. [Google Scholar] [CrossRef]
  38. Wang, Z.; Lai, M.J.; Lu, Z.; Fan, W.; Davulcu, H.; Ye, J. Orthogonal Rank-One Matrix Pursuit for Low Rank Matrix Completion. SIAM J. Sci. Comput. 2014, 37, A488–A514. [Google Scholar] [CrossRef]
  39. Yi, X.; Park, D.; Chen, Y.; Caramanis, C. Fast Algorithms for Robust PCA via Gradient Descent; Neural Information Processing Systems Foundation Inc.: La Jolla, CA, USA, 2016. [Google Scholar]
  40. Sobral, A. BGSLibrary: An OpenCV C++ Background Subtraction Library. In Proceedings of the IX Workshop De Visio Computacional, Rio de Janeiro, Brazil, 3–5 June 2013. [Google Scholar]
  41. Schuhmacher, D.; Vo, B.T.; Vo, B.N. A consistent metric for performance evaluation of multi-object filters. IEEE Trans. Signal Process. 2008, 56, 3447–3457. [Google Scholar] [CrossRef]
Figure 1. Schematic of the proposed method.
Figure 1. Schematic of the proposed method.
Sensors 18 03944 g001
Figure 2. Three typical TIR images. (a) Campus; (b) Parking lot; (c) Community.
Figure 2. Three typical TIR images. (a) Campus; (b) Parking lot; (c) Community.
Sensors 18 03944 g002
Figure 3. The two foreground probability maps of Figure 2a. (a) The original foreground probability map; (b) the foreground probability map with dilation following erosion.
Figure 3. The two foreground probability maps of Figure 2a. (a) The original foreground probability map; (b) the foreground probability map with dilation following erosion.
Sensors 18 03944 g003
Figure 4. The detection results of the 190th frame in Seq 1.
Figure 4. The detection results of the 190th frame in Seq 1.
Sensors 18 03944 g004
Figure 5. The detection results of the 758th frame in Seq. 2.
Figure 5. The detection results of the 758th frame in Seq. 2.
Sensors 18 03944 g005
Figure 6. The detection results of the 173th frame in Seq. 3.
Figure 6. The detection results of the 173th frame in Seq. 3.
Sensors 18 03944 g006
Figure 7. The detection results of the 741rd frame in Seq. 4.
Figure 7. The detection results of the 741rd frame in Seq. 4.
Sensors 18 03944 g007
Figure 8. The cardinality estimates of different trackers.
Figure 8. The cardinality estimates of different trackers.
Sensors 18 03944 g008
Figure 9. The OSPA of different trackers.
Figure 9. The OSPA of different trackers.
Sensors 18 03944 g009
Figure 10. The two typical results of the proposed method. (a) The 20th frame in Seq. 1; (b) the proposed method estimates of (a); (c) the 486th frame in Seq. 4; (b) the proposed method estimates of (c).
Figure 10. The two typical results of the proposed method. (a) The 20th frame in Seq. 1; (b) the proposed method estimates of (a); (c) the 486th frame in Seq. 4; (b) the proposed method estimates of (c).
Sensors 18 03944 g010
Figure 11. The trajectory estimate results.
Figure 11. The trajectory estimate results.
Sensors 18 03944 g011
Table 1. Average recall with different t h d i f f and N.
Table 1. Average recall with different t h d i f f and N.
N1020304050
th diff
200.93150.92190.90570.90820.9241
400.85600.85320.85490.87560.8654
600.73110.78800.80310.81210.8213
800.63430.69960.71460.71080.7288
1000.50830.58040.63070.63160.6423
Table 2. Average precision with different t h d i f f and N.
Table 2. Average precision with different t h d i f f and N.
N1020304050
th diff
200.47100.39370.35570.33610.3158
400.72790.67520.64610.62060.6039
600.83110.82960.81460.80560.8049
800.89100.89490.89310.88890.8853
1000.93160.93040.93740.93490.9338
Table 3. Average FM with different t h d i f f and N.
Table 3. Average FM with different t h d i f f and N.
N1020304050
th diff
200.62210.54820.50610.48470.4665
400.78480.74790.73030.72280.7077
600.77090.80520.80500.80480.8031
800.73580.77960.79090.78490.7964
1000.64800.70620.74980.74770.7552
Table 4. Average evaluation metrics of different p p i x .
Table 4. Average evaluation metrics of different p p i x .
p pix 0.60.70.80.91
recall0.92040.91570.92340.92960.9299
precision0.80850.80470.80200.80370.8020
FM0.85320.84310.85100.85450.8539
Table 5. Average evaluation metrics of different methods.
Table 5. Average evaluation metrics of different methods.
MetricsProposed MethodViBePBASKDESLICOR1MPRPCA-GD
Seq. 1recall0.99960.98750.94340.84770.60900.99920.9996
precision0.94870.84890.77430.82660.84110.57160.6175
FM0.95870.89520.82690.82950.62160.71690.7533
Seq. 2recall0.96250.71270.61320.36600.06790.95980.9621
precision0.82310.84870.85230.75120.06330.678510.8031
FM0.86630.78930.71500.45470.06110.80760.8651
Seq. 3recall0.96560.96320.90070.52760.38690.52380.5778
precision0.84770.49310.65290.54080.13620.78260.7743
FM0.87640.62400.70960.51070.18740.56180.6020
Seq. 4recall0.88470.83110.61500.75240.02100.74770.7786
precision0.79090.71420.87310.63060.02570.81030.7776
FM0.82440.76260.70350.67840.01730.77160.7669
Table 6. Average evaluation metrics for all sequences.
Table 6. Average evaluation metrics for all sequences.
Proposed MethodViBePBASKDESLICOR1MPRPCA-GD
recall0.98390.93320.86190.76820.44350.84490.8714
precision0.87320.61160.81210.59930.30820.78710.7935
FM0.90940.64380.80010.56000.14410.75430.7775
Table 7. Average OSPA values for all sequences.
Table 7. Average OSPA values for all sequences.
Proposed MethodViBe + KalmanPBAS + Kalman
5.0011.139.16

Share and Cite

MDPI and ACS Style

Li, C.; Wang, W. Detection and Tracking of Moving Targets for Thermal Infrared Video Sequences. Sensors 2018, 18, 3944. https://doi.org/10.3390/s18113944

AMA Style

Li C, Wang W. Detection and Tracking of Moving Targets for Thermal Infrared Video Sequences. Sensors. 2018; 18(11):3944. https://doi.org/10.3390/s18113944

Chicago/Turabian Style

Li, Chenming, and Wenguang Wang. 2018. "Detection and Tracking of Moving Targets for Thermal Infrared Video Sequences" Sensors 18, no. 11: 3944. https://doi.org/10.3390/s18113944

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop