Next Article in Journal
A Spatial Decision Support Approach for Flood Vulnerability Analysis in Urban Areas: A Case Study of Tehran
Previous Article in Journal
Rapid Extreme Tropical Precipitation and Flood Inundation Mapping Framework (RETRACE): Initial Testing for the 2021–2022 Malaysia Flood
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Moving Object Tracking Technique Using Few Frames with Feature Map Extraction and Feature Fusion

by
Abeer Abdulaziz AlArfaj
and
Hanan Ahmed Hosni Mahmoud
*
Department of Computer Sciences, College of Computer and Information Sciences, Princess Nourah bint Abdulrahman University, P.O. Box 84428, Riyadh 11671, Saudi Arabia
*
Author to whom correspondence should be addressed.
ISPRS Int. J. Geo-Inf. 2022, 11(7), 379; https://doi.org/10.3390/ijgi11070379
Submission received: 14 May 2022 / Revised: 1 July 2022 / Accepted: 4 July 2022 / Published: 7 July 2022

Abstract

:
Moving object tracking techniques using machine and deep learning require large datasets for neural model training. New strategies need to be invented that utilize smaller data training sizes to realize the impact of large-sized datasets. However, current research does not balance the training data size and neural parameters, which creates the problem of inadequacy of the information provided by the low visual data content for parameter optimization. To enhance the performance of moving object tracking that appears in only a few frames, this research proposes a deep learning model using an abundant encoder–decoder (a high-resolution transformer (HRT) encoder–decoder). An HRT encoder–decoder employs feature map extraction that focuses on high resolution feature maps that are more representative of the moving object. In addition, we employ the proposed HRT encoder–decoder for feature map extraction and fusion to reimburse the few frames that have the visual information. Our extensive experiments on the Pascal DOC19 and MS-DS17 datasets have implied that the HRT encoder–decoder abundant model outperforms those of previous studies involving few frames that include moving objects.

1. Introduction

Tracking of moving objects is an important area of research and a very important task in the field of security automation monitoring systems. The major research of deep learning models for moving object tracking includes region-based CNNs [1] and SSDs, which are single-shot models [2]. Moving object tracking has developed excellent research propositions, namely Overfat [2,3,4] and R-CNN with fast training [4,5,6,7]. These models need a great amount of visual data on a span of many frames for temporal feature maps to enable better performance. Labeling data instances can produce high computational costs as they can be hard to find. This problem has induced the challenges to deep learning, to some extent, and stalled the research of artificial intelligence, which has produced the need to employ a small training data size to track objects in dynamic environments.
Human brains recognize and track moving objects when only sighting a few instances, and we consider that deep neural models should also devise this capability. Assuming a sufficient abundant class of dynamic objects, a different class with few frames can be utilized to track a moving object. As depicted in Figure 1, the training phase is partitioned into two steps: abundant training and scarce training. In abundant training, adequate data training instances are employed to create a feature map space that can characterize the dynamic feature maps of moving objects. In scarce training, a tuning process of the deep learning model allows the new scarce class to be characterized in the feature map space. The authors of [3] considered a few-frames moving object tracking system. It amended the meta-tracking properties by defining the dynamic object vector. Two stages of meta-training process [5] and meta region-based CNN [5,6,7] models were proposed.
The authors of [8] proposed a meta-properties trainer and a property scoring process using the Googlenetv2 model [9] to allow the detection model to adjust to new classes. The feature map learner utilizes enough instances (in the training phase) from abundant labeled data instances to mine meta-feature generalization for future dynamic object classification. Accumulating scores transforms some ancillary instances from the scarce class into vectors that specify the global significance and the high correlation of the computed meta-feature map of the analogous tracked moving object. The process of incorporating the meta-feature map is learned by the feature-map trainer with the score vector computed by the scored procedure. The regression and the prediction information of the dynamic object can be computed. Nevertheless, these models are too solitary for feature map extraction, and many feature representations are unrestricted. Further, the scores are computed by a naive convolution function, which does not take advantage of the current dynamic object classes data.
The proposed HRT encoder–decoder abundant model can tackle this problem. The HRT encoder–decoder can extract all feature representations of a given surveillance video sequence and utilize the support set to achieve feature map fusion and track the dynamic object. The HRT encoder–decoder model extracts the meta feature map of the broad-spectrum dynamic object from the abundant tracking class and then allocates the learned generalized meta feature map to predict other dynamic objects. The proposed model is trained on a large number of labeled instances to mine and compute the selective feature map discriminating vectors that can be employed for some new classes. The proposed model is designed to generate generalization vectors. When utilizing a small number of instances from the new object class to tune the model parameters, the deep learning model can predict the dynamic object of the new scarce class, which is the same for the information extracted from the abundant class transferring to the new class.
The fundamental features of the abundant class represent the common features of both the abundant and new class. Therefore, the basic features of the moving objects have multiple resemblances. For instance, a horse has four tube-shaped legs, as do the new classes of cows and dogs, and so few-frames training resembles a transfer learning concept. For more perception, our model needs two-phase training comprising abundant training with a large dataset that does not have the new class, followed by scarce training which is tuned according to the new class.
In general, the major contributions of this research are discussed as follows:
  • Our research presents an HRT encoder–decoder model that extracts abundant features and an encoder–decoder for feature map fusion to support few-frame moving object tracking.
  • Our research proposes a deep learning model that extracts abundant feature maps and employs parallel temporal and attention procedures.

2. Related Works

2.1. Moving Object Tracking

Machine learning models for moving object tracking utilize surveillance video processing [6,7,8,9,10,11,12]. Moving object tracking utilizes moving object predictions abundant in moving object positions. The current models of moving object tracking comprise of one or many phases for tracking moving objects [13,14,15,16,17], which are defined using the surrounding borders’ geometry computations. The multi-phase model requires the computation of nominee borders with possible moving objects, and then predicts and validates the boundaries of the mined borders [18]. The former model executes a grid technique focused on the moving object to compute prediction parameters to identify the moving object without computing any of the borders. The multi-phase model has a higher performance [19,20,21]. The single phase model is lower compared to the multi-phase model in terms of precision, but it outperforms other models in speed, and it is characterized by a real-time paradigm. Nevertheless, these models have a shortcoming wherein they require more labelled data instances for model training.

2.2. Few-Frames Moving Object Tracking

Few-frames learning is a state-of-the-art model that utilizes only a few frames for object tracking in both training and prediction classification [22,23,24]. The few-frames model includes the few-frames prediction for computing the moving object’s position in surveillance videos. It is used to identify the moving objects with little training data. Usually, we can define this technique as a scarce learning model [25].
The scarce learning model defines the moving object tracking attained from a few analogous data as the starting learning phase. It can be modified to new tracking situations using regression models. The authors of [26] presented a learning algorithm to enhance the performance of the classification of the moving object tracking with little data. The authors of [27] designated moving object information from the moving object province while performing adaptation to augment the cases of few-frames moving object tracking. The model of presenting a preprocessing learning algorithm is usually applied in the moving object tracking paradigm. The notion of a transfer learning technique highly decreases the model training epochs, but it can yield to an overfitting difficulty, which can be solved by the systematic model. Further, while transfer learning with scarce data can converge faster, the classifier still faces an unsatisfactory generalization setting.
Transfer learning allows the model to learn faster. Transfer learning utilizes the pre-learning of similar tasks, and transfer learning is utilized to produce hyper parameters from previous training to attain a worthy model initialization. The transfer learning model presented in [28] syndicates the transfer learning phase and the moving object classification, which can acquire knowledge and fast learning techniques, and train the classifier on how to utilize few data instances. The transfer learning CNN presented in [6] utilizes transfer learning on specific region features and enhances the moving object division through an R-CNN. The authors of [29] use the grouping of moving object tracking and few-frames tracking in an independent transfer learning to offer an approach for adjusting the classifier, and they proposed a higher accuracy tracking system and embraced an optimization technique for few data instances. The authors of [30] proposed a transfer learning initiator to a CNN. Once utilized, given a few surveillance video frames of new moving object, the transfer initiator can yield the classifier impact in the prediction stage of augmented learning of new instances in a feed loop manner. Compared with the Google net model, the authors of [30] utilized the data of many classes, and they utilized the transfer features and adjusted them to obtain fast learning of other classes.

2.3. Attention Model

Encoder–decoder models (E-DM) are always utilized in natural language processors and utilize a transformer model [29,30,31,32]. E-DM models transform the input map M into the following matrices and vectors: support matrix V, key vector K, and value vector VA. E-DMs compute the dot product between V and K to yield the attention score of input instances. This procedure is known as the attention computation, and it is the principal process of encoder–decoder models. The score action of feature map M using the attention scores yields the feature map outputs with specific links between input words. Encoder–decoder models and their deviations have accomplished better results in NPL paradigms. An encoder–decoder model such as the BERT transformer model depicted in [33] trains transformers for unknown translations by mutually restricting the contexts.
Attention models such as the one depicted in [33] employ an encoder–decoder model for spatial data. It was proven that such an encoder–decoder model can enhance the performance of deep learning, especially in computer vision. The authors of [34] introduced a CNN architecture that utilizes an encoder–decoder pair for moving object tracking. In [35], the authors presented an encoder–decoder model to tackle computer vision and accomplished a high accuracy on the data of several surveillance video frame detection inputs. Due to the high accuracy of encoder–decoder models, many computer vision algorithms that depend on the encoder–decoder pair have been introduced. To utilize the encoder–decoder models for few-frames moving object tracking, we propose some alterations to the methodology of the typical encoder–decoder pair model.

3. The Proposed Approach

This research proposes a deep learning model using an abundant encoder–decoder (high resolution transformer (HRT) encoder–decoder). An HRT encoder–decoder employs a feature map extraction that focuses on high-resolution feature maps that are more representative of the moving object. In addition, our research employs the proposed HRT encoder–decoder for feature map extraction and fusion to reimburse the few frames that have the visual information. In the proposed model, we present an abundant class with plenty of data features, while a new scarce class is represented by a small amount of data. Our aim is to utilize the abundant and scarce classes to induce model learning that can predict moving objects in both classes. In Figure 2, the employed double training phase is depicted. The first phase utilizes the preceding information to define the surveillance video frame feature map learned from the abundant class (A). The second phase is the adjusting phase, where tuning of the training of the scarce class occurs to provide adaptation of the neural model to the new moving objects in the scarce class (S). Double data inputs are defined as the support data b and the support data v . The abundant learning phase depicts the definition of the data of the support set to A b and the support data set to A v , which is similarly defined in the new scarce class. If the number of classes in the new class is M and the count of frames in each class is f , the problem is formulated as M -way f -frames moving object tracking.

3.1. The Deep CNN Model

The proposed deep CNN model mainly encompasses a feature map extraction process from the abundant representations, which is utilized to extract feature maps from the support set. Subsequently, the extracted feature representations are depicted into feature map flat vectors, and they are fed to the encoder–decoder input layer. In addition, these vectors are utilized in the auto encoders and decoders to perform feature map fusion. Flat vectors computed from the support and the support sets are utilized to acquire the fusion vectors.
To extract the abundant feature map and decrease the feature map loss of the surveillance video frames, our model employs an abundant feature map training model, as depicted in Figure 3 below.

3.2. The Proposed Support Model

In the proposed support model, the partitioning of the model into multiple parallel phases of self-attention encoders is followed by pooling. Each self-attention encoder is fed by abundant class images and only the last two phases are fed by the scarce class images. Multiple feature sets are extracted from each phase. All feature sets are collected and utilized through a stride fusion model of the feature maps. The resolution from the pooling modules is of a lower resolution than the auto encoder stage. Each parallel phase extracts feature map vectors and then the fusion model performs multi-stride fusion. The feature maps of both the abundant and scarce resolutions are merged, and our model employs stride encoders for the abundant surveillance video frames and down-sampling for the pooling layers. The fusion technique utilizes pixel averaging, and the multiple channels of the resolution feature maps are adjusted to the same value. In addition, to enhance the receptive value, stride convolution computes the feature map vectors in the first phase. In order to improve the focus of the locus of the feature map, we utilize parallel temporal attention at the completion of each phase.
The PTAP process is depicted in Figure 3 and the algorithm is as shown in Algorithm 1. The process encompasses parallel temporal attention threads and is linked via a structural residual model. The attention usually consists of a pooling layer and two 3 × 3 convolution layers, and in between them, an ReLU activation is utilized. The temporal attention process has four convolutions. The attention PTAP centers the attention to decide which channels encompass the main features of the moving dynamic object. The temporal attention process emphasizes the temporal trajectory and detects which frame includes the main data of the moving dynamic object. The detail of the process is depicted as follows:
M a p F = I C · ψ ( M a p T A + F C A )
M a p S A = C P ( C G ( I c ) )
M a p C A = C u ( μ ( C D ( M a p P ( I c ) ) ) )
M a p F defines the feature map of the output and the input ( I C ). The activation functions ψ and μ indicate Sigmoid and ReLu activation functions. C P , C D , and C u define convolution functions, and M a p P indicates the maximum pooling average value. M a p T A represents the temporal attention output value and M a p C A depicts the channel attention output.
Our attention model employs a parallel design to learn surveillance video frame complex feature representations. The parallel temporal attention modules can compute the channel features correlation at each point and adjust them to improve the representative ability of feature representations mapping.

3.3. The Encoder–Decoder Model

The encoder–decoder model was introduced by Google [35,36,37,38,39,40]. An HRT encoder–decoder utilizes a self-attention method to compute the feature maps in a parallelized mechanism from the surveillance video input. To keep the input correlation, the model utilizes position-coding to compute the location coordinates. Consequently, the encoder–decoder model can guarantee the correlation between the previous and subsequent data; but, due to the parallel nature of the input, the model training period is decreased. The encoder–decoder model has a transformer structure encoder. When extracting the feature map, the parallel input is fed to the encoder for the correlation computation and other data feature maps are acquired and decoded.

3.3.1. The Encoder Structure

The key module of the encoder in the encoder–decoder model is the self-attention method. To compute the attention vector, the three inputs of the location, namely I1, I2, I3, are utilized as depicted in Algorithm 1.
Algorithm 1: The Encoder Framework
1. Compute the correlation in the input. The correlation is computed by the dot product, which is to compute the dot product for the vectors in I 1 and each vector in I 2 . The specific formula is:
Corr   = I 1 I 2 T
2. The computed correlation is divided by the parameter d to alleviate the gradient in the learning phase, as depicted in Equation (5):
Corr   =   Corr / d
where d defines the distribution parameter of the classifier softmax and denotes the model degeneration learning curve.
3. Change the normalized correlation vector into a value in the range of zero and one using the softmax classifier. The correlation is converted into a probability matrix Z with values in the range of zero and one, as follows:
Z = Attension ( I , J ) = Softmax   ( IJ T / d )
4. Compute the value of the dot product of Z and K.
X = Z K
The purpose of accumulating a residual instance RES is to avoid the degradation in the deep neural model of the training model. Degradation implies that with the accumulation of the layers’ number in the deep neural model, the loss yields to attain saturation and the layer count increases.
Normalization can speed up the learning process and enhance the learning curve stability. Nevertheless, normalization has to solve the small size of the data. The normalization layer is linked to the input size, and if the input is small, it will face high interference. The mean and variance of the input will yield a misrepresentation of the data distribution. This can yield to usage of a large amount of memory, as well as an extended learning time. The learning phase may fail due to the static gradient path. In this case, we can utilize channel normalization, which splits the channel into sub-channels and computes inside the batch. The computation does not depend on the sub-channel size, and the performance can be more stabilized in bigger batches. Channel normalization can prevent batch normalization problems. For surveillance video frames with batch sizes of M, G, H, and C, channel normalization defines the channels into sub-channels and computes averages and standard deviations in each sub-channel, forcing each layer input to follow the range of zero to one distribution, which resolves the covariance problem of relocation and speeds up the model’s convergence. This is depicted as follows:
r = I E [ I ] SD [ I ] + ϵ p 1 + p 2
where I is the input, r is the normalized input, E [ I ] is the expected value, SD [ I ] is the standard deviation, p 1 and p 2 are the training parameters, and ϵ is the threshold that stops the denominator from reaching zero.

3.3.2. The Decoder Structure

In the encoder–decoder model the decoder structure transmits the support map to the support feature map. The support set vector and the support vector are fed as inputs to the encoder’s I and J . Concurrently, we subdue the background external to the support moving object, and we use the label of the support vector as the training phase input with the computed support vector from the equation B M. Then, we compute the transformed feature map using the attention equation of Z B I ( B M ) . The computation is depicted as follows:
I c h a n n e l = channel . Norm ( Z B I ^ ( B M ,   J , I ) + I )
Compared with I , the feature-map-enhanced I c h a n n e l groups various moving object maps from the support feature map I to enhance its value.
The input of the combined feature maps form a feed forward network (FF) with ‘Avoid Connection’. Its importance lies in the process of the ReLU layer. The feature map vector is extracted by the attention process to a feature map adaptation, which enhances the model’s expressiveness. The FF model is a double/multi-layer perceptron model (D-MLP), which has a fully connected (FC) layer and an ReLU activation layer, which is employed to each location distinctly. The computation is as follows:
F F ( O ) = R e L U ( O C 1 + a 1 ) C 2 + a 2
Here, O is the output from the prior layer, and
where ,   C 1 S D m × D f ,
C 2 S D f × D m ,
a 1 S D f , and
a 2 S D m .
These are all the hyper-parameters of the training phase. The value of the parameter D f is higher than the value of D m . After the transitory stage of the FF network, we utilize the Acc and the channel normalization processes.

4. Experiments

In this section, we will compare and test the HRT encoder–decoder model through a model simulation. In this article, an encoder–decoder model is utilized to identify few-frames moving object tracking. The experiments are depicted in the following subsections.

4.1. Datasets

We utilized public data for moving object tracking surveillance to train and test our model. There were two datasets: DOC19 and DS17. The data set description of the HRT encoder–decoder model is depicted in [12].

4.1.1. The Dataset DOC19

We utilized the DS17 and DOC19 datasets for model training using 12,000 video frames. The validation process used video frames from both datasets (5300). The training dataset chose abundant classes, while the prediction process used new instances. The abundant classes contained many labelled surveillance video frame data, and the new class had few surveillance video frames. For the N-class and M-frames surveillance tracking process, we defined the new class as N classes, and each single class had M video frames with tagged labels. In the beginning, we accomplished model training on the abundant classes to get a primary model score, and at the next stage, we performed a fine tuning of the model on the new class. In the new class, we accumulated the moving object in the abundant class so that the trained encoder–decoder model could identify both the new and the abundant classes. To avoid the non-generality of the model tracking process, we split the dataset into three subsets to train and test the model. In each subset, for the 22-class, five classes were selected as the new classes, and the other classes were utilized as the abundant-class data. For each subset, we took 3, 7, and 9 for the K parameter of the new class for training and validation. When evaluating the datasets, we utilized the mean accuracy of the new class to test. When the join and difference ratio between the result and the true label was higher than 0.5, then the result was correct, that is, JD50.

4.1.2. The Dataset DS17

The DS17dataset enclosed rich classes and a large number of video frames. It will be utilized for testing surveillance video frame object tracking. In the paradigm of the moving object tracking process, DS17 included 76 different classes with 10,000 video frames for training and 5000 video frames for validation. We selected 18 classes as the new class sets and the remaining classes were the abundant class set.

4.1.3. Training Process

The simulation environment of our experiment was a TX208 GPU with 64 GB of memory. It was executed using Python on Linux sun stations and utilized deep learning PyTorch to construct the encoder–decoder models. The model parameter gradient utilized stochastic descent computation with an energy of 0.8, and score tuning was defined as being of value equal to 0.0005, with batch defined as 32 in size. In addition, the training surveillance video frames were handled by a model of horizontal, vertical overturning, and color exposure to increase the training data size.

4.2. Comparison

4.2.1. Results on the Dataset DOC19

In this section, we describe the experimental results. Table 1 illustrates the results of the model when trained on the dataset DOC19 of the HRT encoder–decoder model on the new class. We also compare the results with current single-phase models, such as SPD [41], Meta Googlenet [42], and Det [43]. The proposed HRT encoder–decoder predictor performs at higher tracking results when the number of video frames of the new class is high. In the first subset, we developed 1.4% more than the others in five frames, 3.3% points more than the others in seven frames, and 1.3% over the others in eleven frames.

4.2.2. Results on the Dataset DS17

Compared with the DOC19 dataset, DS17 show more complications in moving object tracking processes due to the DS17 set having more video frames. We performed the training of the model with 60 abundant classes of DS17, and then we performed the fine-tuning process of the model when the frames were 13 or 23, separately. The results are depicted in Table 2. Our proposed model outperforms previous models. When the frame number was 13, our model enhanced the performance by 8.1% at JD45: 90, and for a frame number of 23, our model enhanced the performance by 9.3% at JD45: 90. The results are illustrated in Figure 4 and Figure 5.

4.3. Ablation Experiments

The ablation results are decisive in utilizing the encoder–decoder HRT for infusion and in using an abundant extraction model. The ablation training and testing were performed on the DOC19 dataset, and the frame number was defined to be 7, with the abundant and the new class splitting of the dataset.
To test the temporal attention in the abundant CNN, we accumulated the ablation results of these processes. We performed the experiments in the CNN abundant model. As depicted in Table 3, the experimental results are enhanced when temporal attention was accumulated. When channel attention was accumulated in the model, the results were enhanced more. Consequently, we found that the accumulative receptive parameter of the model and a higher defined score of the feature map using the attention method is highly operative.
Using the attention method in the encoder–decoder model shows an enhancement of the encoder–decoder in the computer vision model in [33]. Table 4 depicts the ablation results. The fusion of the abundant set feature map vector with the scarce vector shows a better performance. In addition, when a facade is utilized to substitute the preceding temporal location in the decoding phase of the encoder–decoder model, the moving object tracking ablation result of the model can also be enhanced.
After the ablation simulation on the feature map computational model and encoder–decoder, the best arrangement of the models is found to perform the ablation computation on the abundant feature map model and encoder–decoder. Table 5 depicts the results after 200 epochs of training. We also confirmed the impact of the accumulated procedures with our training models. Through this accumulation, the scarce model results can be highly enhanced for moving object tracking.

5. Conclusions

In this research, we introduced an HRT encoder–decoder model to recognize few-frames moving object tracking. In the model, we utilize an abundant feature map extraction model to extract model feature maps, as well as an attention encoder–decoder to infuse the support set feature maps and the support feature maps. An effective predictor is proposed by merging the abundant model and the encoder–decoder model to be applied to new scarce instances. The experimental results proved that our proposed HRT encoder–decoder model performs better than the preceding classifiers when the number of video frames is higher than three. We also confirmed the impact of the accumulated procedures with our training models. Through this accumulation, the scarce model results can be highly enhanced for moving object tracking.

Author Contributions

Conceptualization, Hanan Ahmed Hosni Mahmoud and Abeer Abdulaziz AlArfaj; methodology, Hanan Ahmed Hosni Mahmoud; software, Hanan Ahmed Hosni Mahmoud; validation, Hanan Ahmed Hosni Mahmoud and Abeer Abdulaziz AlArfaj formal analysis, Hanan Ahmed Hosni Mahmoud; investigation, Hanan Ahmed Hosni Mahmoud; resources, Hanan Ahmed Hosni Mahmoud; data curation, Hanan Ahmed Hosni Mahmoud; writing—original draft preparation, Hanan Ahmed Hosni Mahmoud; writing—review and editing, Hanan Ahmed Hosni Mahmoud; visualization, Hanan Ahmed Hosni Mahmoud; supervision, Hanan Ahmed Hosni Mahmoud; project administration, Hanan Ahmed Hosni Mahmoud; funding acquisition, Abeer Abdulaziz AlArfaj. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Princess Nourah bint Abdulrahman University Researchers Supporting Project number (PNURSP2022R113), Princess Nourah bint Abdulrahman University, Riyadh, Saudi Arabia.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Availble upon request.

Acknowledgments

We would like to thank the following for funding our project: Princess Nourah bint Abdulrahman University Researchers Supporting Project number (PNURSP2022R113), Princess Nourah bint Abdulrahman University, Riyadh, Saudi Arabia.

Conflicts of Interest

The authors declare that they have no conflict of interest to report regarding the present study.

References

  1. Engel, J.; Schops, T.; Cremers, D. Direct monocular non-alluded SLAM. Comput. Vis. 2019, 8, 834–849. [Google Scholar]
  2. Murrtal, R.; Tardos, J. SCR: An open-source SLAM system for monocular stereo and RGB-D cameras. IEEE Trans. Robot. 2021, 33, 1255–1262. [Google Scholar] [CrossRef] [Green Version]
  3. Bescos, A.; Facil, J.; Neira, J. DynaS: Tracking mapping and in painting in dynamic scenes. IEEE Robot. Auton. Lett. 2019, 3, 4076–4083. [Google Scholar] [CrossRef] [Green Version]
  4. Ronzoni, D.; Olmi, R.; Fantuzzi, C. AGV global localization using indistinguishable artificial landmarks. In Proceedings of the IEEE Conference on Robotics, Cairo, Egypt, 31 May 2020; pp. 287–292. [Google Scholar]
  5. Hafez, R.; David, J. SLAM2: A SLAM system for monocular stereo and RGB-D cameras. IEEE Trans. Robot. 2020, 33, 1255–1262. [Google Scholar]
  6. Mime, J.; Bayoun, D. LSD: Large static direct monocular model. Comput. Vis. 2020, 7, 83–89. [Google Scholar]
  7. Ahmed, M.; Cremers, D. Indirect deep learning odometer model. IEEE Trans. Trans. Pattern Anal. Mach. Intell. 2019, 4, 61–65. [Google Scholar]
  8. Engel, J.; Koltun, V.; Cremers, D. Direct sparse odometer. IEEE Trans. Pattern Anal. Mach. Intell. 2018, 40, 611–625. [Google Scholar] [CrossRef]
  9. Collins, R.; Zhou, X.; Teh, S.K. An open source tracking testbed and evaluation web site. In Proceedings of the IEEE Workshop Performance Evaluation Tracking Surveillance, Paris, France, 20 March 2020; pp. 17–24. [Google Scholar]
  10. Xu, H.; Yang, M.; Wang, X.; Yang, Q. Magnetic sensing system design for intelligent vehicle guidance. IEEE/ASME Trans. Mechatron. 2020, 15, 652–656. [Google Scholar]
  11. Loevsky, I.; Shimshoni, I. Reliable and efficient landmark-abundantd localization for mobile robots. Robot. Auton. Syst. 2021, 58, 520–528. [Google Scholar] [CrossRef]
  12. Geiger, A.; Lenz, P.; Stiller, C.; Urtasun, R. Vision meets robotics: The VID dataset. J. Robot. Reason. 2021, 32, 123–127. [Google Scholar]
  13. Fisher, R. The MOVSD4 surveillance ground-truth data sets. In Proceedings of the IEEE Workshop Performing Evaluation Tracking Surveillance, London, UK, 29 January 2019; pp. 12–17. [Google Scholar]
  14. Fuentes, J.; Ascencio, J.; Mancha, J. Visual simultaneous localization and mapping: A survey. Artif. Intell. Rev. 2019, 43, 55–81. [Google Scholar] [CrossRef]
  15. Saputra, M.; Markham, A.; Trigoni, N. Visual SLAM and structure from motion in dynamic environments: A survey. ACM Comput. Surv. 2020, 51, 37. [Google Scholar] [CrossRef]
  16. Cadena, C.; Carlone, L.; Carrillo, H.; Latif, Y. Past present and future of simultaneous localization and mapping: Toward the robust-perception age. IEEE Trans. Robot. 2016, 32, 1309–1332. [Google Scholar] [CrossRef] [Green Version]
  17. Wang, Y.; Lin, M.; Ju, R. Visual SLAM and moving-object detection for a small-size humanoid robot. Adv. Robot. Syst. 2021, 7, 133–143. [Google Scholar] [CrossRef] [Green Version]
  18. Kundu, A.; Krishna, K.; Sivaswamy, J. Moving object detection by multi-view geometric techniques from a single camera mounted robot. In Proceedings of the IEEE/RSJ Conference on Robotics, Lafayette, LA, USA, 4–8 November 2019; pp. 436–441. [Google Scholar]
  19. Li, S.; Lee, D. RGB-D SLAM in dynamic environments using static point weighting. IEEE Robot. Autosomes 2020, 2, 223–230. [Google Scholar] [CrossRef]
  20. Tan, W.; Liu, H.; Bao, H. Robust monocular SLAM in dynamic environments. In Proceedings of the IEEE Mixed Augmented Reality, Athens, Greeece, 27 May 2019; pp. 209–218. [Google Scholar]
  21. Carion, N.; Massa, F.; Synnaeve, G.; Usunier, N.; Kirillov, A.; Zagoruyko, S. End-to-End moving object tracking with Encoder-decoders. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020; pp. 213–229. [Google Scholar]
  22. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An surveillance video frame is worth 16x16 words: Encoder-decoders for surveillance video frame recognition at scale. arXiv 2020, arXiv:2010.11929. [Google Scholar]
  23. Sun, K.; Xiao, B.; Liu, D.; Wang, J.D. Deep abundant map learning for human pose estimation. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 5693–5703. [Google Scholar]
  24. Park, J.; Woo, S.; Lee, J.-Y.; Kweon, I.S. Bam: Bottleneck attention module. arXiv 2018, arXiv:1807.06514. [Google Scholar]
  25. Woo, S.; Park, J.; Lee, J.Y.; Kweon, I.S. Cbam: Convolutional block attention module. In Proceedings of the European Conference on Computer Vision, Munich, Germany, 8–14 September 2018; pp. 3–19. [Google Scholar]
  26. Fischler, M.; Bolles, R. Random sample consensus: A paradigm for model fitting with applications to image analysis and automated cartography. Commun. ACM 2021, 24, 381–395. [Google Scholar] [CrossRef]
  27. Klein, G.; Murray, D. Parallel tracking and mapping for small AR workspaces. ACM Mix. Augment. Real. 2017, 2, 225–234. [Google Scholar]
  28. Alcantarilla, P.; Yebes, J.; Almazan, J.; Bergasa, L. On combining visual SLAM and dense scene flow to increase the robustness of localization and mapping in dynamic environments. In Proceedings of the IEEE Conference on Robotics, Alexandria, Egypt, 29–31 October 2019; pp. 190–197. [Google Scholar]
  29. Giordano, D.; Murabito, F.; Spampinato, C. Superpixel-abundantd video object segmentation using perceptual organization and location prior. Comput. Pattern Recognit. 2020, 6, 484–489. [Google Scholar]
  30. Sun, Y.; Liu, M.; Meng, M. Improving RGB-D SLAM in dynamic environments: A motion removal approach. Robot. Syst. 2021, 8, 110–122. [Google Scholar] [CrossRef]
  31. Yu, C.; Liu, Z.; Wei, Q. DS-SLAM: A semantic visual SLAM towards dynamic environments. In Proceedings of the IEEE/RSJ Conference on Robotics, Venus, Italy, 20 September 2018; pp. 1168–1174. [Google Scholar]
  32. Cheng, Y.; Meng, M. Semantic mapping in dynamic environments. Robotica 2020, 38, 256–270. [Google Scholar] [CrossRef]
  33. Hirschmuller, H. Stereo processing by semiglobal matching and mutual information. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 30, 328–341. [Google Scholar] [CrossRef] [PubMed]
  34. Lowe, D. Distinctive image features from scale-invariant key points. J. Comput. 2020, 6, 91–110. [Google Scholar]
  35. Bay, H.; Ess, A.; Gool, L.V. SURF: Speeded up robust features. Proc. Conf. Computation. Vis. 2021, 3, 346–359. [Google Scholar]
  36. Rosten, E.; Porter, R.; Drummond, T. Faster and better: A machine learning approach to corner detection. IEEE Trans. Pattern Anal. Mach. 2021, 32, 105–119. [Google Scholar] [CrossRef] [Green Version]
  37. Achanta, R.; Sässtrunk, S. SLIC super pixels compared to state-of-the-art super pixel methods. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 4, 227–234. [Google Scholar]
  38. Sturm, J.; Cremers, D. A benchmark for the evaluation of RGB-D SLAM systems. In Proceedings of the IEEE/RSJ Conference of Intelligent Robots, Beijing, China, 27 May 2017; pp. 573–580. [Google Scholar]
  39. Kerl, C.; Sturm, J.; Cremers, D. Robust odometry estimation for RGB-D cameras. In Proceedings of the IEEE Conference on Robots, New Delhi, India, 14–18 October 2019; pp. 3748–3754. [Google Scholar]
  40. Sun, Y.; Meng, M. Motion removal for reliable RGB-D SLAM in dynamic environments. Robot. Auton. Syst. 2018, 10, 115–128. [Google Scholar] [CrossRef]
  41. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. SPD: Attention is all you need. In Advances in Neural Information Processing Systems; The MIT Press: Los Angeles, CA, USA, 2017; pp. 5998–6008. [Google Scholar]
  42. Devlin, J.; Chang, M.-W.; Lee, K.; Toutanova, K. Meta Googlenet: Transfer learning of deep bidirectional encoder-decoders. Trans. Pattern Anal. Mach. Intell. 2020, 2, 22–34. [Google Scholar]
  43. Wu, H.P.; Liu, Y.L.; Wang, J.W. Review of Det dataset in Deep Learning. Comput. Mater. Contin. 2020, 63, 1309–1321. [Google Scholar]
Figure 1. Abundant training versus scarce training of the model.
Figure 1. Abundant training versus scarce training of the model.
Ijgi 11 00379 g001
Figure 2. The proposed HRT encoder–decoder model: a support phase and a support learning phase. The HRT encoder–decoder has three processes: support extraction module, auto encoder, and encoder–decoder.
Figure 2. The proposed HRT encoder–decoder model: a support phase and a support learning phase. The HRT encoder–decoder has three processes: support extraction module, auto encoder, and encoder–decoder.
Ijgi 11 00379 g002
Figure 3. Feature map extraction model consisting of multiple parallel phases followed by a fusion model.
Figure 3. Feature map extraction model consisting of multiple parallel phases followed by a fusion model.
Ijgi 11 00379 g003
Figure 4. Average accuracy for k = 13.
Figure 4. Average accuracy for k = 13.
Ijgi 11 00379 g004
Figure 5. Average accuracy for k = 23.
Figure 5. Average accuracy for k = 23.
Ijgi 11 00379 g005
Table 1. Results using DOC19 where the first subset used five frames, seven frames, and eleven frames, and the JD50 result was obtained.
Table 1. Results using DOC19 where the first subset used five frames, seven frames, and eleven frames, and the JD50 result was obtained.
New Subset 1New Subset 2New Subset 3
Frames351135103510
SPD12.429.138.55.015.731.015.027.336.3
Meta Googlenet26.733.947.222.730.140.528.442.845.9
Det28.935.048.825.930.641.527.941.942.9
Our HRT model29.437.749.927.836.847.528.944.249.2
Table 2. Results of the proposed HRT encoder–decoder model using the DS17 data set with frame numbers of 13 and 23 compared to other models in few-frames moving object tracking.
Table 2. Results of the proposed HRT encoder–decoder model using the DS17 data set with frame numbers of 13 and 23 compared to other models in few-frames moving object tracking.
Average Accuracy %
Frame NumberModelJD45: 900.50.75
k = 13SPD848181
Meta Googlenet828886
Det838681
Our HRT encoder–decoder949697
k = 23SPD878681
Meta Googlenet818786
Det838781
Our HRT encoder–decoder959698.4
Table 3. The ablation results with the feature map extraction model accumulated with the channel and temporal attention.
Table 3. The ablation results with the feature map extraction model accumulated with the channel and temporal attention.
Temporal AttentionChannel AttentionAccumulated Temporal and Channel AttentionJD50
0.531
0.542
0.546
0.558
Table 4. Ablation results of facade and attention fused with the encoder–decoder.
Table 4. Ablation results of facade and attention fused with the encoder–decoder.
FacadeAttentionJD50
0.6006
0.6220
0.6642
0.6462
Table 5. Ablation results of abundant feature map extraction model, encoder–decoder, and scarce model utilized in the proposed HRT model.
Table 5. Ablation results of abundant feature map extraction model, encoder–decoder, and scarce model utilized in the proposed HRT model.
Scarce ModelAbundant ModelCombined Abundant
and Scarce Model
JD50
0.8545
0.8202
0.8548
0.8482
0.8550
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Abdulaziz AlArfaj, A.; Mahmoud, H.A.H. A Moving Object Tracking Technique Using Few Frames with Feature Map Extraction and Feature Fusion. ISPRS Int. J. Geo-Inf. 2022, 11, 379. https://doi.org/10.3390/ijgi11070379

AMA Style

Abdulaziz AlArfaj A, Mahmoud HAH. A Moving Object Tracking Technique Using Few Frames with Feature Map Extraction and Feature Fusion. ISPRS International Journal of Geo-Information. 2022; 11(7):379. https://doi.org/10.3390/ijgi11070379

Chicago/Turabian Style

Abdulaziz AlArfaj, Abeer, and Hanan Ahmed Hosni Mahmoud. 2022. "A Moving Object Tracking Technique Using Few Frames with Feature Map Extraction and Feature Fusion" ISPRS International Journal of Geo-Information 11, no. 7: 379. https://doi.org/10.3390/ijgi11070379

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop