Next Article in Journal
Feasibility of Using Elastic Wave Velocity Monitoring for Early Warning of Rainfall-Induced Slope Failure
Next Article in Special Issue
Assessment of Spatiotemporal Fusion Algorithms for Planet and Worldview Images
Previous Article in Journal
Infrared Hollow Optical Fiber Probe for Localized Carbon Dioxide Measurement in Respiratory Tracts
Previous Article in Special Issue
Convolutional Neural Network-Based Shadow Detection in Images Using Visible Light Camera Sensor
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Computationally Efficient Automatic Coast Mode Target Tracking Based on Occlusion Awareness in Infrared Images

1
Agency for Defense Development, P.O. Box 35, Yuseong, Daejeon 34186, Korea
2
Department of Electronic Engineering, Yeungnam University, 280 Daehak-ro, Gyeongsan, Gyeongbuk 38541, Korea
3
Department of Electrical Engineering, KAIST, 291 Daehak-ro, Yuseong-gu, Daejeon 34141, Korea
*
Author to whom correspondence should be addressed.
Sensors 2018, 18(4), 996; https://doi.org/10.3390/s18040996
Submission received: 1 February 2018 / Revised: 24 March 2018 / Accepted: 25 March 2018 / Published: 27 March 2018
(This article belongs to the Special Issue Sensors Signal Processing and Visual Computing)

Abstract

:
This paper proposes the automatic coast mode tracking of centroid trackers for infrared images to overcome the target occlusion status. The centroid tracking method, using only the brightness information of an image, is still widely used in infrared imaging tracking systems because it is difficult to extract meaningful features from infrared images. However, centroid trackers are likely to lose the track because they are highly vulnerable to screened status by the clutter or background. Coast mode, one of the tracking modes, maintains the servo slew rate with the tracking rate right before the loss of track. The proposed automatic coast mode tracking method makes decisions regarding entering coast mode by the prediction of target occlusion and tries to re-lock the target and resume the tracking after blind time. This algorithm comprises three steps. The first step is the prediction process of the occlusion by checking both matters which have target-likelihood brightness and which may screen the target despite different brightness. The second step is the process making inertial tracking commands to the servo. The last step is the process of re-locking a target based on the target modeling of histogram ratio. The effectiveness of the proposed algorithm is addressed by presenting experimental results based on computer simulation with various test imagery sequences compared to published tracking algorithms. The proposed algorithm is tested under a real environment with a naval electro-optical tracking system (EOTS) and airborne EO/IR system.

1. Introduction

Automatic tracking of an electro-optical system refers to keeping a target at the center of the images by driving the sensor’s LOS (line of sight) according to the result of a video tracker that calculates a change in the target’s position using acquired images. As shown in Figure 1, a video tracker detects a target’s location from input images and sends the pixel error (difference between target’s location and center in the image) to the servo controller. Then, the servo controller drives gimbals as a mount of tracking error to locate the target at the center of images. Iteration of that process keeps a target in the sensor’s line of sight. These kinds of automatic tracking systems have a variety of uses in military and security systems because they provide automotive operation and a target’s accurate position simultaneously.
Although image tracking has been the focus of many studies over the world, there are still many challenging issues such as overcoming target’s occlusions, appearance changes, significant motions, background clutters, etc. Among the challenges in the image tracking area, the occlusion problem is one of the most crucial problems.
In this paper, a novel automatic coast mode tracking method is proposed to prevent from target loss against the target occlusion, which is one of the functions for improving the operability of centroid tracker for infrared images, considering the computational cost of fast moving target tracking. The proposed algorithm consists of three steps. The first step is to make a decision to enter the coast mode that estimates when the target occlusion occurs. The second step is to perform memory tracking using a prediction result of the target’s location rather than image tracking result during the target’s blind time. The last step is to re-lock the target (or re-acquisition target) after the target occlusion is released. This study shows effective performance according to the results based on computer simulations with various test imagery compared with several published tracking algorithms. It is tested under a real environment with an electro-optical tracking system (EOTS) mounted on ship and airborne EO/IR (electro-optic/infrared) system.
Section 2 introduces the background of coast mode tracking and the proposed occlusion aware–coast mode tracking (OA-CMT) method is explained in Section 3. In Section 4, experimental results through the simulation and real EOTS dataset validate the feasibility of the proposed method in various occlusion scenarios. Section 5 concludes this paper.

2. Background of Coast Mode Tracking

A video tracker finds a target’s location in the acquired images from imaging sensors and calculates tracking error using image center information. There is a problem of tracking loss affected by abrupt obstacles, because the tracking result is computed depending on the brightness of each pixels of image. There have been various attempts to solve the occlusion problem in tracking. For example, previous methods using appearance models [1,2,3], adaptive appearance modeling by statistical analysis [4,5,6], and template matching of target models [7,8] showed good performance for limited situations, but it is easy to degrade tracking performance because blind updates in long-term occlusion status causes contamination of appearance models. Multi-camera-based methods [9,10,11] were also proposed, but these methods require a complex multi-camera setup with high cost, which is not appropriate for a moving camera platform such as EOTS. Patch matching-based methods [12,13,14] can widely handle not only the occlusion problem but can also overcome appearance changes. However, they can deal with a short-duration of partial occlusion and fail at severe occlusions lasting for a long time. Probabilistic approaches [15,16,17,18] and L1 trackers [19,20] have been proposed several times. State-of-the-art trackers, upgraded correlation trackers [21] and neural network-based tracking methods [22,23,24], have been presented, which show outstanding tracking performance in multiple target tracking, cluttered environments, changing of target’s shape, and occlusion. However, these trackers require an on-line or off-line training process with an amount of data sets of targets to accommodate various situations, meaning they are not suitable for tracking systems that must track arbitrary targets what the operator chooses.
The aforementioned techniques have two limitations in EOTS applications. One is the computational cost for real time processing on embedded system. They need more than 30 frames per second to track a fast-moving target. Algorithms using complex calculation are useless for on-line systems. The other problem is that those kinds of tracking method are not applicable very well in the case of infrared images because they deal with features of imagery with clear edge or color information. Infrared sensors, recording radiant intensity that is emitted by a targets’ thermal energy, are superior to visible cameras in detecting and tracking small objects at long distance and can be used at night time, which leads to the wide use of infrared sensors in military applications. However, it is difficult to extract the feature points for tracking because the shapes of targets are usually shown blurred and blinked due to atmospheric scattering and air turbulence between the target and sensor. Therefore, it is better to use target intensity information for target detection and tracking in infrared images. The targets to track are normally mapped to high intensity regions compared to the surrounding background.
The proposed algorithm can overcome the target occlusion status using centroid trackers for infrared images where the extraction of meaningful feature is difficult. The centroid tracking method, using only the brightness information of an image, is still used widely in infrared tracking systems because of its computational simplicity.
The coast mode refers to one of the tracking modes preventing tracking loss against a target’s occlusion [25]. As shown in Figure 2, the image tracker is halted for a while and turned to a memory tracking status when the reliability of tracking becomes lower or the prediction of target occlusion event happens on normal tracking status. The memory tracking refers to the status that a servo drives a sensor’s line of sight in the predicted direction and speed from movement data is measured in normal tracking status before the occlusion. Using predicted results rather than image tracking results during occlusion status, which obtain meaningless results from the image tracker, helps to increase of the possibility of resuming tracking after long-term occlusion while some trackers suffer from maintaining image tracking over long periods of occlusion. After the blind time, it tries to find the target to resume video tracking. When the target can be detected and becomes reliable enough to be re-locked, image tracking is started. Otherwise, the video tracker stops if it fails to re-lock on the target after a certain period of time.
In general, the coast tracking mode could be divided into forced coast tracking, automatic coast tracking [25], and gun coast tracking [26] according to the system requirement. Forced coast tracking runs by user’s decision and operation; it is also referred to as manual coast tracking mode. When the operator selects forced coast mode, it starts memory tracking without image tracker until he locks on the target to track. The automatic coast tracking means the automatic process of coast tracking for user’s convenience, so the decision algorithm is required to enter coast status and use the re-locking algorithm with image processing. The gun coast tracking is used to avoid tracking loss caused by gun firing. It enters coast status automatically with gun fire command input. In this case, the target re-locking process is tried after the fixed time that is expected for disappearing gun smoke.

3. Proposed OA-CMT Based Infrared Target Tracker

The proposed occlusion-aware coast mode tracking (OA-CMT) is performed in the order shown in Figure 3. In a newly inputted frame image, it predicts if the target is occluded by an obstacle through image processing around the target being tracked. When there is no occlusion, it continues image tracking and collects information of the brightness, shape and motion of the target. If any obstacle is detected around the target, it stops image tracking, switches to memory tracking status, predicting the target’s location using historically filtered data, and then tries to reacquire the target. When the occlusion is ended and the target candidate is detected, the image tracking is resumed if the image tracking result is reliable. Otherwise, it will continue maintaining memory tracking and attempting to reacquire the target.

3.1. Prediction of Target’s Obstruction

The awareness of target occlusion is made by two steps. In the first step, the “target likelihood obstacle check” predicts a target’s occlusion by checking the brightness similarity of the background in the direction of the target’s movement: if the background brightness around a target is similar to that of the target’s, the centroid tracker would miss the target. In the second step, “background screening check” predicts the loss of the tracking situation where the background appears which can hide the target. If any target likelihood obstacles and background likelihood obstacles are detected, it stops the centroid tracking for the target, turns the tracking status to the coast mode and starts memory tracking.
The target likelihood obstacle check consists of the steps of searching for targeted candidates with similar brightness, carrying the history of the candidates over frames, and determining the occlusion alarm for target likelihood, as shown in Figure 4.
During the tracking, it is more likely to be affected by an object which has similar brightness to the target and is located close to the target. The first step, searching for candidates, is to detect the objects of particular brightness that are likely to be classified as a target by the centroid tracker. For the efficiency of computation time, it is necessary to set an appropriate searching area around the target and divide it into several blocks. The brightness comparison for screening the target likelihood area is performed on a block related to the target size (set as a quarter of the target’s size) rather than on a pixel-by-pixel basis. The searching area is set to a multiple of the block’s size considering target’s speed and moving direction. Then, it classifies an area as a “bright block” if the count of pixels brighter than the threshold value is more than 50% of the block area. Then, it registers a candidate by clustering the adjacent “bright blocks” with the information such as the size, location, average brightness, and the “true distance (Dtrue)”. True distance means the distance between the outer surfaces of target and obstacle, as shown in Figure 5. With the similar triangles rule, Dmin (minimum distance from center of the target to outer surface of obstacle) and (Dmin − Dtrue) have a proportional relation with their x-components, Dmin and HalfTgX (half of target size), then Dtrue could be approximately calculated as Equation (1).
D true = D min × ( D minX Half TgX ) D minX ,
The second step is to manage the histories of each candidate over frames to predict whether candidates affect the centroid tracking process. The matching of candidates between the current frame and previous frame is performed by comparing the registered information with the effective distance. In this step, the process is carried out with three cases: (1) the observed candidate having no history (that means it is first observed candidate); (2) having history but not being an observed candidate (that means a disappeared candidate); (3) observed candidate having history. If a candidate has history, the effective distance that the candidate is expected to move in the next frame is set to manage the change trend of the candidate over frames and used for matching between the observed candidate and history in the next frame.
The occlusion alarm is determined by the candidate’s approaching distance to the target. First, it computes average velocity for each of the candidates, calculates the “estimated distance (Dest)” that reflects the prediction of the obstacle’s movement on the Dtrue, such as in Equation (2), then checks if the estimated distance is smaller than a distance threshold value. At this time, the distance threshold value is set to be proportional to the block size considering the block unit operation in the first step. It makes an alarm of target likelihood occlusion when the candidate is present and the estimated distance is smaller than the threshold value.
D est = D true + ScaleFactor × ( D true D true ,   history ) ,
Figure 6 shows the results of the target likelihood obstacle check in the test image sequence which is made to for computer simulation. (a–c) are partial views of a test image sequence where a 16 × 16 pixel-sized target approaches the obstacle with an average brightness of target. (d–f) are the results of searching candidates for (a–c) with magnifying interested area. In (d–f), the square box represents the search area, and a white spot means a bright block. The operational interval of blocks could be guessed from the distance between spots. There are two clusters of spots: the smaller one is the target and the bigger one is the obstacle. The calculation results of estimated distance are 16, 11, 6 for (d–f), respectively. An occlusion alarm is made at (f).
The background screening check is a process to determine if the track is difficult to keep because an obstacle located between the IR camera and the target obscures the view of the camera even though its brightness is not similar the target. It consists of three steps: the first step is the brightness change check of four guard gates and four background gates as shown in Figure 7. The second step is to make a pre-alarm with a synchronization check of sequential occurrence of brightness changes between guard gate and background gate. The last step is to determine the occlusion alarm for background screening with pre-alarm and size tracking window. Figure 7 presents the procedure of the background screening check.
During the first step, brightness change around the target is determined using χ2 distribution for their brightness. First of all, four background gates and four guard gates about four directions are set around the target as shown in Figure 8. The cumulative distribution function (CDF) of average (μ) and standard deviation (σ) for brightness on each gate over frames are computed. As shown in Figure 9, a region of α sized area is set as polluted criteria that brightness change is detected; otherwise, the region of (1–α) sized area is set normal criteria. As represented in Equation (3), if the chi-square distribution value is smaller than ϵ , it is determined as a normal situation, otherwise the gate is determined to be polluted by an obstacle in the case of greater than ϵ . To solve the scintillation problem of IR imagery, applying a normalization factor considering global intensity is recommended, as in Equation (3).
p ( X normal   status ) = p ( [ X / NF μ ] T σ 1 [ X / NF μ ] σ ϵ ) p ( X pollusion   status ) = p ( [ X / NF μ ] T σ 1 [ X / NF μ ] σ ϵ ) ,
In the second step, a pre-alarm is made when the synchronicity of polluted gates with the sequence is admitted. There is an observed sequence because the first guard gate is polluted and then the background gate in the same direction will be polluted within several frames when an obstacle approaches to the target. Synchronization of gate pollution is determined more accurately when considering a target’s moving direction and speed.
The last step is to make a final decision of the background screening check regarding the size reduction of the target gate of the same direction with gate pollution under pre-alarm situation.
Figure 10 shows the result of the background screening check for a test image sequence which is made to perform a computer simulation, as in the case of a target likelihood obstacle check but with a darker obstacle than target. (a–d) show the process of gates’ pollution over frames by obstacle, while (e–h) present the status of image tracking. With this sequence, the brightness of the right side of the target changes as the target approaches the obstacle, as in Figure 11. χ value increases when the intensity mean of the frame (x-axis) of the background gate definitely changes compared to intensity mean over sequence (μ). As the target is being hidden behind the obstacle, the gate size becomes smaller than its average size. With those clues, it can be seen that background screening alarm occurs in Figure 10h, as the target tracking gate disappears because the obscuring alarm makes image tracking stop.

3.2. Memory Tracking Using Tracking Filter

If a target’s obstruction is predicted with the process described in the previous chapter, it stops the image tracker and performs memory tracking with inertia until the target is re-locked. In the memory tracking status, the image tracker cannot output the pixel error, and the target’s location must be predicted with tracking filter. Of course, it is possible to acquire more accurate location of the target using advanced tracking filters such as the interactive multiple model (IMM) filter [27] or particle filter with three-dimensional information [28] if it can obtain the distance to the target. However in this study, α-β-γ filter [29] is simply selected to get a position for re-locking process without range information.
The α-β-γ filter is a simplified form of the Kalman filter for estimation, data smoothing and control applications. It is suitable for predicting a target’s location in memory tracking because it keeps the effect of Kalman filter in a time-invariant state without a detailed system model. Starting from a one-dimensional, linear, time-invariant, ideal target motion model described in Equation (4), where y(k) is the target state vector (composed of location x(k), velocity v(k), and acceleration a(k)) at time k, w(k) is the unknown target maneuver, and their transition matrix φ(k), ψ(k). Normally, measurement z(k) is composed by sum of position and noise n(k) as in Equation (5), the target prediction and correction are solved like Equation (6), and prediction Equation (7).
y ( k + 1 ) = φ · y ( k ) + ψ · w ( k ) [ x ( k + 1 ) v ( k + 1 ) a ( k + 1 ) ] = [ 1 T T 2 / 2 0 1 T 0 0 1 ] [ x ( k ) v ( k ) a ( k ) ] + [ T 2 / 2 T 1 ] w ( k ) ,
z ( k ) = h · y ( k ) + n ( k ) z ( k ) = [ 1 0 0 ] [ x ( k ) v ( k ) a ( k ) ] + n ( k ) ,
y ˇ ( k + 1 | k + 1 ) = y ˇ ( k + 1 | k ) + K ( k + 1 ) [ z ( k + 1 ) h · y ˇ ( k + 1 | k ) ,
y ˇ ( k + 1 | k ) = φ · y ˇ ( k | k ) ,
K = [ Kx Kv Ka ] = [ α β / T γ / T 2 ]
K, steady state Kalman gain, is related to the α-β-γ parameters as shown in Equation (8). The optimal α-β-γ relationships are previously reported [29].
A tracking filter is implemented using the tracking index parameter of time period T = 1/30 s, target maneuverability = 3 g (m/s2), and measurement noise = 0.2 m. The tracking result with test imagery sequences is shown in Figure 12. While the target is under occlusion, the location of the target is predicted by the α-β-γ filter. The dotted square represents the prediction of the target’s position regarded as being hidden by an obstacle in Figure 12. The performance of the filter can be verified from the result that the dotted square appears superimposed on the location of the target after blind time.

3.3. Target’s Re-Locking

When the target is seen in the image after blind time, the process of re-locking target is performed in order to resume image tracking. The proposed method is designed using the target brightness value, which is the main characteristic of infrared images.
Spatial or temporal filters are often applied for target detection in the case of dealing with intensity information. However, these methods are likely to fail to detect targets when the brightness of the target is changed by the AGC (automatic gain control) function of infrared cameras. Therefore, in the proposed study, the computing histogram ratio is used between the target and background.
Figure 13 shows the procedure of the proposed target re-locking algorithm. First, the target model is established by gathering information of the target during image tracking. When a decision of occlusion is made, it starts screening the blocks which have a high probability of being the target within the search area divided into blocks. The probability of each block is calculated by the sum of the histogram ratio modeled in the previous step, and blocks having a probability higher than the threshold value are selected as the target candidates. Next, the selected candidates are scored by comparison to target model, and the highest scored candidate is picked as the target.
In the first step of the target re-locking algorithm, on-line training of the target model is performed during image tracking status. This method calculates the histogram ratio between the target gate and four background gates (except guard gates in Figure 8), while the original histogram back-projection method [30] is computed for the entire image. It causes not only heavy computation but also less accuracy because the background histogram of the whole image may change according to the target’s movement.
Equation (9) [31] shows the formula for target model with a histogram ratio of target model Hr(i), target’s histogram value Ht(i), sum of background gates’ histogram value, for gray level i. In Equation (9), a high histogram ratio of grey level i means a high probability that level I belongs to the target [32]. To get a more accurate target model, Hr(i) must be averaged over frames, especially in the case of using a histogram stretched image.
H r ( i ) = H t ( i ) H t ( i ) + H t ( i ) + 1 ,
Other information such as target size, location, shape, and intensity variance are collected and statistically modeled in this step to be applied for the scoring process.
A transition to the coast mode occurs by determining occlusion, and the image process for detecting target candidates is activated until image tracking resumes. The searching candidate process works within the searching area considering target’s moving speed and direction. For reduced computation time and de-noising effect, it is performed by the unit of blocks referred in the previous section, “prediction of target’s obstruction”.
For each block, the average is computed of the extracted value of the histogram ratio model corresponding to the brightness of each pixel within the block. This average value presents the probability of how similar the block is to the target. Figure 14b shows the target probabilities of the test image, a higher intensity block presents a higher probability of being the target.
After the probability values are acquired about all blocks, clustering is performed for the adjacent blocks which are screened by the probability threshold. These clusters are regarded as target candidates: bigger or smaller clusters are excluded compared to the target’s size.
In the final step of the target re-locking algorithm, the score is calculated to determine the target to track and decide whether to resume tracking. For the target candidates, the similarity with the target is scored by comparison to the target model including size, intensity variation, aspect ratio, and target probability (i.e., average value of histogram ratio over pixels) that are statically collected in image tracking status. The candidate with the highest score is selected as the target, and the tracking reliability is checked to make a decision whether to switching to image tracking status.
Figure 15 shows the result of the target’s re-locking process. Figure 15a shows that the re-locking algorithm is being executed during blind time. After blind time, the target appears, the target gate is located on the target as a result of the re-locking algorithm. Figure 15c shows that image tracking is resumed through the checking track ability.

4. Experimental Results

In this section, to verify the performance of the proposed method, OA-CMT, computer simulations have been carried out on several infrared and visible images. Table 1 shows the list of the test imagery sequences including descriptions about sensor types, targets, and obstacles.
Figure 16, Figure 17, Figure 18, Figure 19, Figure 20, Figure 21 and Figure 22 show the results of the computer simulations compared to other published trackers dealing with target occlusion. In each figure, (a–d) are selected scenes showing meaningful results among sequences, including tracking result of Tracking Learning Detection (TLD) [33] (red box), L1 tracker using Accelerated Proximal Gradient (L1-APG) [20] (violet box), Sparsity-Based Collaborative Model (SCM) [18] (cyan box), and OA-CMT (green box and green dot during coast status).
Each tracking result also includes tracking trajectories of vertical and horizontal locations over frames. In the case of OA-CMT, there are no big differences between the predicted results (presented as green dots) and re-locked locations during target occlusion. It is not desirable for EOTS that the tracking result includes abrupt changes of target’s location as in Figure 18, Figure 19 and Figure 22, because the servo controller may cause difficulties in driving gimbals according to the tracking result.
To verify applications for real time embedded systems, computational cost must be considered. With the computer platform Intel® Core™ i5 CPU @2.27GHz, manufactured by HP (Palo Alto, CA, USA), computation costs were measured. As shown in Table 2, OA-CMT is fastest compared with other trackers for all test images. In addition, since the computation time required to perform the proposed algorithm depends on the image size, the specification of the embedded board can be determined according to the image resolution of the sensor.

5. Conclusions

This paper proposes an occlusion aware-coast mode tracking algorithm for infrared imagers. OA-CMT includes the decision algorithm of predicting target occlusion with image processing, the tracking filter to output predictions of target position in blind time, and post-blind time automatic re-locking for user’s convenience. The proposed algorithm helps the sensor driving unit to continue its movement during blind time caused by target occlusion.
The prediction of target obstruction is designed by checking the occlusion both of target likelihood brightness and background screening around the target considering the target’s size, moving direction and speed. For memory tracking, the α-β-γ filter is used to predict the target’s position. The target re-locking algorithm is proposed using target modeling of histogram ratio between target and background.
Computer simulations with various test imagery sequences are carried out and their results are compared to other occlusion-related published tracking algorithms, TLD, L1-APG, and SCM. Computational costs are also presented for considering real time embedded systems. After implementation on a video tracking board, the proposed algorithm is undergoing validation tests in real environment for applications for naval EOTS (shown in Figure 23) and airborne EO/IR system (shown in Figure 24) After implementation on the video tracking board of EOTS, the performance of the proposed algorithm is validated through tests under real environment.

Acknowledgments

This study is supported in part by Agency for Defense Development with NCUAS (Next generation Corps Unmanned Aircraft System) EO/IR project, by the National Research Foundation of Korea within the Ministry of Science, ICT and Future Planning through the Korean Government under Grant NRF-2017R1A2A2A05001400, and by the ICT R&D program of MSIP/IITP 2016-0-00563, Research on Adaptive Machine Learning Technology Development for Intelligent Autonomous Digital Companion.

Author Contributions

The contributions were distributed between authors as follows: Sohyun Kim wrote the text of the manuscript, programmed OA-CMT. Gwang-Il Jang proposed operational concept and designed base structure of OA-CMT. Sungho Kim and Junmo Kim performed in-depth discussion of the related literature and experiment results.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Senior, A.; Hampapur, A.; Tian, Y.; Brown, L.; Pankanti, S.; Bolle, R. Appearance models for occlusion handling. Image Vis. Comput. 2006, 24, 1233–1243. [Google Scholar] [CrossRef]
  2. Nguyen, H.T.; Smeulders, A.W.M. Fast occluded object tracking by a robust appearance filter. IEEE Trans. Pattern Anal. Mach. Intell. 2004, 26, 1099–1104. [Google Scholar] [CrossRef] [PubMed]
  3. Huan, Y.; Essa, I. Tracking multiple object through occlusions. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, San Diego, CA, USA, 20–25 June 2005. [Google Scholar]
  4. Han, B.; Davis, L. On-line density-based appearance modeling for object tracking. In Proceedings of the 10th IEEE International Conference on Computer Vision, Beijing, China, 17–21 October 2005. [Google Scholar]
  5. Jepson, A.; Fleet, D.; El-Maraghi, T. Robust online appearance models for visual tracking. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Kauai, HI, USA, 8–14 December 2001. [Google Scholar]
  6. Ross, D.; Lim, J.; Yang, M. Adaptive probabilistic visual tracking with incremental subspace update. In Proceedings of the 8th European Conference on Computer Vision, Prague, Czech Republic, 11–14 May 2004. [Google Scholar]
  7. Choi, W.; Jung, J.; Park, D.; Choi, B.; Choi, S. Automatic tracking system with target classification. In Proceedings of the SPIE Defense, Security, and Sensing, Orlando, FL, USA, 13–17 April 2009; Volume 7335. [Google Scholar]
  8. Pan, J.; Hu, B. Robust Occlusion Handling in Object Tracking. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Minneapolis, MN, USA, 17–22 June 2007. [Google Scholar]
  9. Chang, T.; Gong, S.; Ong, E. Tracking multiple people under occlusion using multiple cameras. In Proceedings of the 11th British Machine Vision Conference, Bristol, UK, 11–14 September 2000. [Google Scholar]
  10. Dockstander, S.L.; Tekalp, A.M. Multiple camera fusion for multi-object tracking. In Proceedings of the 2001 IEEE Workshop on Multi-Object Tracking, Vancouver, BC, Canada, 8 July 2001; pp. 95–102. [Google Scholar]
  11. Fleuret, F.; Berclaz, J.; Lengagne, R.; Fua, P. Multi-camera people tracking with a probabilistic occupancy map. IEEE Trans. Pattern Anal. Mach. Intell. 2008, 30, 267–282. [Google Scholar] [CrossRef] [PubMed]
  12. Yao, R.; Shi, Q.; Shen, C.; Zhang, Y.; van den Hengel, A. Part-based visual tracking with online latent structural learning. In Proceedings of the 2013 IEEE Conference on Computer Vision and Pattern Recognition, Portland, OR, USA, 23–28 June 2013. [Google Scholar]
  13. Yang, M.; Wu, Y.; Hua, G. Spatial selection for attentional visual tracking. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Minneapolis, MN, USA, 17–22 June 2007. [Google Scholar]
  14. Zhang, T.; Jia, K.; Xu, C.; Ma, Y.; Ahuja, N. Partial occlusion handling for visual tracking via robust part matching. In Proceedings of the 2014 IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014. [Google Scholar]
  15. Isard, M.; Blake, A. Condensation-Conditional density propagation for visual tracking. Int. J. Comput. Vis. 1998, 29, 5–28. [Google Scholar] [CrossRef]
  16. Isard, M.; Blake, A. Contour tracking by stochastic propagation of conditional density. In Proceedings of the 4th European Conference on Computer Vision, Cambridge, UK, 15–18 April 1996. [Google Scholar]
  17. Doucet, A.; Freitas, N.; Gordon, N. Sequential Monte Carlo Methods in Practice; Springer: Berlin, Germany, 2001; pp. 3–14. [Google Scholar]
  18. Zhong, W.; Liu, H.; Yang, M.-H. Robust object tracking via sparsity-based collaborative model. In Proceedings of the 2012 IEEE Conference on Computer Vision and Pattern Recognition, Providence, RI, USA, 16–21 June 2012. [Google Scholar]
  19. Mei, X.; Ling, G.; Wu, Y.; Blasch, E.; Bai, L. Minimum error bounded efficient L1 tracker with occlusion detection. In Proceedings of the 2011 IEEE Conference on Computer Vision and Pattern Recognition, Colorado Springs, CO, USA, 20–25 June 2011. [Google Scholar]
  20. Mei, X.; Blasch, E.P. Efficient minimum error bounded particle resampling L1 tracker with occlusion detection. IEEE Trans. Image Process. 2013, 22, 2661–2675. [Google Scholar] [PubMed]
  21. Zhang, B.; Luan, S.; Chen, C.; Han, J.; Wang, W.; Perina, A.; Sha, L. Latent constrained correlation filter. IEEE Trans. Image Process. 2018, 27, 1038–1048. [Google Scholar] [CrossRef]
  22. Held, D.; Thrun, S.; Savarese, S. Learning to track at 100 fps with deep regression networks. In Proceedings of the 14th European Conference, Amsterdam, The Netherlands, 11–14 October 2016. [Google Scholar]
  23. Sadeghian, A.; Alahi, A.; Savarese, S. Tracking the untrackable: Learning to track multiple cues with long-term dependencies. In Proceedings of the 2017 IEEE Computer Vision and Pattern Recognition, Venice, Italy, 22–29 October 2017. [Google Scholar]
  24. Hong, S.; You, T.; Kwak, S.; Han, B. Online tracking by learning discriminative saliency map with convolutional neural network. In Proceedings of the International Conference on Machine Learning, Lille, France, 6–11 July 2015. [Google Scholar]
  25. Kim, S.; Jang, G.; Kwon, K.H.; Jung, J.H. Design of autocoast tracking algorithm by the prediction of target occlusion and its on-based implementation. J. Korea Inst. Mil. Sci. Technol. 2009, 12, 354–359. [Google Scholar]
  26. Kim, S.; Jang, G. Decision method of tracking possibility under the firing flame influence in IR image. In Proceedings of the Institute of Electronics Engineers of Korea Fall Conference, Daejeon, Korea, 26–27 November 2011. [Google Scholar]
  27. Bar-Shalom, Y.; Li, X.R.; Kirubarajan, T. Estimation with Applications to Tracking and Navigation; Wiley: New York, NY, USA, 2001. [Google Scholar]
  28. Gustafsson, F.; Gunnarsson, F.; Bergman, N.; Forssell, U.; Jansson, J.; Karlsson, R.; Nordlund, P.-J. Particle filters for positioning, navigation and tracking. IEEE Trans. Signal Process. 2002, 50, 425–437. [Google Scholar] [CrossRef]
  29. Kalata, P.R. A generalized parameter for α-β and α-β-γ target trackers. IEEE Trans. Aerosp. Electron. Syst. 1984, AES-20, 174–182. [Google Scholar] [CrossRef]
  30. Lee, J.H.; Lee, W.H.; Jeong, D.S. Object tracking method using back projection of multiple color histogram models. Circuits Syst. 2003, 2, 668–671. [Google Scholar]
  31. Kim, S.; Kim, J. Design of re-locking algorithm using target modeling of histogram ratio during coast tracking mode in infrared image. In Proceedings of the Unmanned/Unattended Sensors and Sensor Networks IX, Edinburgh, UK, 24–27 September 2012. [Google Scholar]
  32. Wu, J.; Mao, S.; Wang, X.; Zhang, T. Ship target detection and tracking in cluttered infrared imagery. Opt. Eng. 2011, 50, 057207. [Google Scholar] [CrossRef]
  33. Kalal, Z.; Matas, J.; Mikolajczyk, K. P-N Learning: Bootstrapping Binary Classifiers by Structural Constraints. In Proceedings of the 2010 IEEE Conference on Computer Vision and Pattern Recognition, San Francisco, CA, USA, 13–18 June 2010. [Google Scholar]
Figure 1. Automatic tracking procedure of an electro-optical system.
Figure 1. Automatic tracking procedure of an electro-optical system.
Sensors 18 00996 g001
Figure 2. State diagram of the coast mode tracking process.
Figure 2. State diagram of the coast mode tracking process.
Sensors 18 00996 g002
Figure 3. Flowchart of the coast mode tracking.
Figure 3. Flowchart of the coast mode tracking.
Sensors 18 00996 g003
Figure 4. Procedure of target likelihood brightness check.
Figure 4. Procedure of target likelihood brightness check.
Sensors 18 00996 g004
Figure 5. Calculating true distance (Dtrue) between the target and an obstacle.
Figure 5. Calculating true distance (Dtrue) between the target and an obstacle.
Sensors 18 00996 g005
Figure 6. Result of target likelihood obstacle check in test image sequence (ac) are test image sequences, and (df) show the results of searching candidates for (ac) with magnifying interested area. In (df), the square box represents the search area spots, meaning a bright block. Calculation result of estimated distance is (d) 16, (e) 11, (f) 6.
Figure 6. Result of target likelihood obstacle check in test image sequence (ac) are test image sequences, and (df) show the results of searching candidates for (ac) with magnifying interested area. In (df), the square box represents the search area spots, meaning a bright block. Calculation result of estimated distance is (d) 16, (e) 11, (f) 6.
Sensors 18 00996 g006
Figure 7. Procedure of background screening check.
Figure 7. Procedure of background screening check.
Sensors 18 00996 g007
Figure 8. Shape of background gates and guard gates. There are four gates for each direction: up, down, left, and right side around target. Both background gate and guard gate are 1/4 size of target gate.
Figure 8. Shape of background gates and guard gates. There are four gates for each direction: up, down, left, and right side around target. Both background gate and guard gate are 1/4 size of target gate.
Sensors 18 00996 g008
Figure 9. Cumulative distribution function of χ2 distribution.
Figure 9. Cumulative distribution function of χ2 distribution.
Sensors 18 00996 g009
Figure 10. Result of background screening obstacle check in test image sequence. (ad) show the process of gates’ pollution over frames by obstacle; (eh) present the status of image tracking. In (h), the background screening alarm finally occurs.
Figure 10. Result of background screening obstacle check in test image sequence. (ad) show the process of gates’ pollution over frames by obstacle; (eh) present the status of image tracking. In (h), the background screening alarm finally occurs.
Sensors 18 00996 g010
Figure 11. Brightness change histories of the right side gate among four-directional background gates are shown when the target moves in the right direction. χ value (chi) increases as the intensity mean of the frame mean (right gate frame mean) definitely changes compared to intensity mean over the sequence (right gate sequence mean). As the target is being hidden behind an obstacle, gate size became smaller than its average size.
Figure 11. Brightness change histories of the right side gate among four-directional background gates are shown when the target moves in the right direction. χ value (chi) increases as the intensity mean of the frame mean (right gate frame mean) definitely changes compared to intensity mean over the sequence (right gate sequence mean). As the target is being hidden behind an obstacle, gate size became smaller than its average size.
Sensors 18 00996 g011
Figure 12. Result of memory tracking using α-β-γ filter in test imagery sequences. Dotted square is the prediction of target’s position.
Figure 12. Result of memory tracking using α-β-γ filter in test imagery sequences. Dotted square is the prediction of target’s position.
Sensors 18 00996 g012
Figure 13. Procedure of target re-locking.
Figure 13. Procedure of target re-locking.
Sensors 18 00996 g013
Figure 14. Result of screening target candidates.
Figure 14. Result of screening target candidates.
Sensors 18 00996 g014
Figure 15. Result of target re-locking. During blind time, the re-locking algorithm is being executed as shown in (a). It is shown that the target is re-locked after blind time in (b). In (c), image tracking is resumed through checking ability of tracking.
Figure 15. Result of target re-locking. During blind time, the re-locking algorithm is being executed as shown in (a). It is shown that the target is re-locked after blind time in (b). In (c), image tracking is resumed through checking ability of tracking.
Sensors 18 00996 g015
Figure 16. Computer simulation result of imagery ID 1. (a) Normal tracking status (original image); (b) magnifying interested area; (c) TLD loses track while the car is partially hidden by a street lamp; (d) OA-CMT turns to memory tracking status because the car is hidden by a bush; (e) tracking trajectories of each tracker. The green dot says OA-CMT turns to memory tracking status while other trackers are tracking the wrong position.
Figure 16. Computer simulation result of imagery ID 1. (a) Normal tracking status (original image); (b) magnifying interested area; (c) TLD loses track while the car is partially hidden by a street lamp; (d) OA-CMT turns to memory tracking status because the car is hidden by a bush; (e) tracking trajectories of each tracker. The green dot says OA-CMT turns to memory tracking status while other trackers are tracking the wrong position.
Sensors 18 00996 g016
Figure 17. Computer simulation result of imagery ID 2. (a) Normal tracking status (original image); (b) magnifying interested area; (c) L1-APG loses track while the car is partially hidden by a street lamp while OA-CMT turns to memory tracking status and will re-lock successfully after several frames; (d) TLD tracks wrong target;(e) tracking trajectories of each tracker. Around the 600th frame, TLD presents a very different position because of tracking the wrong target (at the time of (d)). After the 700th frame, the car is hidden by trees, OA-CMT turns to memory tracking status, while other trackers give wandering positions.
Figure 17. Computer simulation result of imagery ID 2. (a) Normal tracking status (original image); (b) magnifying interested area; (c) L1-APG loses track while the car is partially hidden by a street lamp while OA-CMT turns to memory tracking status and will re-lock successfully after several frames; (d) TLD tracks wrong target;(e) tracking trajectories of each tracker. Around the 600th frame, TLD presents a very different position because of tracking the wrong target (at the time of (d)). After the 700th frame, the car is hidden by trees, OA-CMT turns to memory tracking status, while other trackers give wandering positions.
Sensors 18 00996 g017
Figure 18. Computer simulation result of imagery ID 3. (a) Normal tracking status (original image); (b) magnifying interested area; (c) OA-CMT turns to memory tracking status when the tank is hidden by a human; (d) TLD, L1-APG and OA-CMT re-lock target successfully; (e) tracking trajectories of each trackers During blind time, tracking results of TLD are hopping because it tracks the human. SCM lost tracking with the changing shape of the target when tank turned its direction.
Figure 18. Computer simulation result of imagery ID 3. (a) Normal tracking status (original image); (b) magnifying interested area; (c) OA-CMT turns to memory tracking status when the tank is hidden by a human; (d) TLD, L1-APG and OA-CMT re-lock target successfully; (e) tracking trajectories of each trackers During blind time, tracking results of TLD are hopping because it tracks the human. SCM lost tracking with the changing shape of the target when tank turned its direction.
Sensors 18 00996 g018
Figure 19. Computer simulation result of imagery ID 4. (a) Normal tracking status (original image); (b) magnifying interested area; (c) OA-CMT turns to memory tracking status as the car is hidden by a building (d) TLD and OA-CMT re-lock target successfully; (e) tracking trajectories of each tracker. Predicted position of target of OA-CMT has no big difference with the position of the target after blind time.
Figure 19. Computer simulation result of imagery ID 4. (a) Normal tracking status (original image); (b) magnifying interested area; (c) OA-CMT turns to memory tracking status as the car is hidden by a building (d) TLD and OA-CMT re-lock target successfully; (e) tracking trajectories of each tracker. Predicted position of target of OA-CMT has no big difference with the position of the target after blind time.
Sensors 18 00996 g019
Figure 20. Computer simulation result of imagery ID 5. (a) Normal tracking status (original image); (b) OA-CMT turns to memory tracking status when the tank is hidden by trees; (c) other trackers are wandering; (d) TLD, SCM and OA-CMT re-lock target successfully after blind time; (e) tracking trajectories of each tracker. TLD follows the wrong target during blind time.
Figure 20. Computer simulation result of imagery ID 5. (a) Normal tracking status (original image); (b) OA-CMT turns to memory tracking status when the tank is hidden by trees; (c) other trackers are wandering; (d) TLD, SCM and OA-CMT re-lock target successfully after blind time; (e) tracking trajectories of each tracker. TLD follows the wrong target during blind time.
Sensors 18 00996 g020
Figure 21. Computer simulation result of imagery ID 6. (a) normal tracking status (original image); (b) magnifying interested area; (c) OA-CMT turns to memory tracking status while the bus is hidden by the overpass (bridge) but other trackers are tracking the bridge; (d) OA-CMT re-locks target successfully, TLD re-locks the back side of the target while other tracker loses track; (e) tracking trajectories of each tracker. After blind time (from 250th frame) tracking positions of TLD are hopping because it confused the target with the other bus.
Figure 21. Computer simulation result of imagery ID 6. (a) normal tracking status (original image); (b) magnifying interested area; (c) OA-CMT turns to memory tracking status while the bus is hidden by the overpass (bridge) but other trackers are tracking the bridge; (d) OA-CMT re-locks target successfully, TLD re-locks the back side of the target while other tracker loses track; (e) tracking trajectories of each tracker. After blind time (from 250th frame) tracking positions of TLD are hopping because it confused the target with the other bus.
Sensors 18 00996 g021
Figure 22. Computer simulation result of imagery ID 7. (a) Normal tracking status (original image); (b) magnifying interested area; (c) OA-CMT turns to memory tracking status as the car is hidden by the traffic sign; (d) L1-APG and OA-CMT re-locks target successfully while TLD and SCM lost the target; (e) tracking trajectories of each tracker.
Figure 22. Computer simulation result of imagery ID 7. (a) Normal tracking status (original image); (b) magnifying interested area; (c) OA-CMT turns to memory tracking status as the car is hidden by the traffic sign; (d) L1-APG and OA-CMT re-locks target successfully while TLD and SCM lost the target; (e) tracking trajectories of each tracker.
Sensors 18 00996 g022
Figure 23. Result of application for naval EOTS: (a) shape of naval EOTS (bd) operational images of testing coast mode tracking.
Figure 23. Result of application for naval EOTS: (a) shape of naval EOTS (bd) operational images of testing coast mode tracking.
Sensors 18 00996 g023
Figure 24. Result of application for airborne EO/IR: (a) shape of airborne EO/IR (bd) operational images of testing coast mode tracking.
Figure 24. Result of application for airborne EO/IR: (a) shape of airborne EO/IR (bd) operational images of testing coast mode tracking.
Sensors 18 00996 g024
Table 1. Lists of the test imagery sequences.
Table 1. Lists of the test imagery sequences.
IDImagery SequencesSensorObstacle TypeDescription
1 Sensors 18 00996 i001IRBackground screeningCar screened by bush
2 Sensors 18 00996 i002IRBackground screeningTruck screened by street lamp & trees
3 Sensors 18 00996 i003IRTarget likelihood obstacleTank screened by human
4 Sensors 18 00996 i004IRBackground screeningCar screened by building
5 Sensors 18 00996 i005EOBackground obstacleCar screened by trees
6 Sensors 18 00996 i006EOBackground screeningBus screened by trees & bridge
7 Sensors 18 00996 i007IRTarget likelihood obstacleCar screened by traffic sign
Table 2. Comparison of computational cost.
Table 2. Comparison of computational cost.
IDImage SizeTarget SizeFrames Per Second
TLDL1-APGSCMOA-CMT (Proposed)
1720 × 480(336 frames)56 × 487.050.030.9842
2720 × 480(831 frames)28 × 246.390.021.0930.81
3640 × 480(130 frames)36 × 143.882.542.8971.5
4640 × 480(283 frames)10 × 810.53.421.4871
5320 × 240(300 frames)90 × 409.872.922.18149.5
61280 × 1024(310 frames)94 × 262.132.641.186.6
71280 × 1024(480 frames)24 × 161.370.31.436.67

Share and Cite

MDPI and ACS Style

Kim, S.; Jang, G.-I.; Kim, S.; Kim, J. Computationally Efficient Automatic Coast Mode Target Tracking Based on Occlusion Awareness in Infrared Images. Sensors 2018, 18, 996. https://doi.org/10.3390/s18040996

AMA Style

Kim S, Jang G-I, Kim S, Kim J. Computationally Efficient Automatic Coast Mode Target Tracking Based on Occlusion Awareness in Infrared Images. Sensors. 2018; 18(4):996. https://doi.org/10.3390/s18040996

Chicago/Turabian Style

Kim, Sohyun, Gwang-Il Jang, Sungho Kim, and Junmo Kim. 2018. "Computationally Efficient Automatic Coast Mode Target Tracking Based on Occlusion Awareness in Infrared Images" Sensors 18, no. 4: 996. https://doi.org/10.3390/s18040996

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop