You are currently on the new version of our website. Access the old version .
SensorsSensors
  • Article
  • Open Access

20 March 2023

Online Learning-Based Hybrid Tracking Method for Unmanned Aerial Vehicles

,
,
and
1
Department of Multimedia Engineering, Hanbat National University, Daejeon 34158, Republic of Korea
2
Electronics and Telecommunications Research Institute, Daejeon 34129, Republic of Korea
*
Author to whom correspondence should be addressed.
This article belongs to the Section Sensor Networks

Abstract

Tracking unmanned aerial vehicles (UAVs) in outdoor scenes poses significant challenges due to their dynamic motion, diverse sizes, and changes in appearance. This paper proposes an efficient hybrid tracking method for UAVs, comprising a detector, tracker, and integrator. The integrator combines detection and tracking, and updates the target’s features online while tracking, thereby addressing the aforementioned challenges. The online update mechanism ensures robust tracking by handling object deformation, diverse types of UAVs, and changes in background. We conducted experiments on custom and public UAV datasets to train the deep learning-based detector and evaluate the tracking methods, including the commonly used UAV123 and UAVL datasets, to demonstrate generalizability. The experimental results show the effectiveness and robustness of our proposed method under challenging conditions, such as out-of-view and low-resolution scenarios, and demonstrate its performance in UAV detection tasks.

1. Introduction

With the recent development of drone technologies, the practical applications of unmanned aerial vehicles (UAVs), also known as drones, are becoming increasingly diversified [1]. However, the diversity of using drones can incur negative effects, since drones have multidirectional purposes. For example, unauthorized use of UAVs, e.g., hovering over airports, military facilities, and private compounds, can pose threats related to security and privacy. An accident at these locations could lead to serious disasters, endangering both human life and property. To mitigate the adverse effects of such incidents, governments regulate UAV operations through civil aeronautics laws. However, it is challenging to restrict all uncontrollable UAVs [1]. To address the unauthorized use of UAVs, various systems using radar, radio frequency signals, and images to detect, identify, and track illegal UAVs have been actively investigated [2]. This paper focuses on the issue of single object tracking in images captured from the ground or sky to ensure the surveillance of unauthorized UAVs.
Visual object tracking in images is a fundamental computer vision task, e.g., automatic driving, augmented reality, and visual surveillance. Despite significant advancements in deep learning-based object tracking in recent years, there are still several challenges that need to be addressed, such as dealing with distractors caused by similar objects, and changes in the appearance of the tracked object [3,4]. In addition, tracking can easily drift to the background due to occlusions, target objects that are out of view, and the abrupt appearance of objects. The small size of objects makes tracking extremely difficult because of the limited amount of available visual information. Tracking UAVs in outdoor environments is particularly challenging for the following reasons: (1) wide-area surveillance, including the sky and ground, because of UAVs’ wide range of activities; (2) the diverse sizes and appearances of UAVs; (3) low image resolution because moving objects can occupy a small area; (4) significant distance between the camera and the moving objects; (5) large interframe disparity caused by fast UAV and camera motion; and (6) uncontrollable environmental conditions, e.g., illumination changes and weather conditions.
Many tracking methods have been proposed in recent years to handle these challenging problems. For example, one approach integrates tracking and detection, where the key is how to determine the output from the tracker and detector. Wang et al. [5] combined the tracker and detector with a linear combination. In addition, Saribas et al. [6] proposed a mechanism to switch between the tracker and detector. This mechanism compares the Euclidean distances and intersection-over-union (IoU) ratio between the detected bounding boxes and the latest output of the tracker. However, these integration methods [5,6] focus on the location and size of the bounding boxes obtained from the tracker and detector, rather than the target’s appearance. As a result, these methods are limited in their ability to accommodate object deformation. Baptista et al. [7] proposed a surveillance system in which the deep convolutional ResNet-50 model performs object classification on tracked aerial objects to detect aerial targets. Although the pretrained classifier realizes reasonable performance in terms of tracking and classifying the target, the pretrained model may be limited in terms of generalizability due to a lack of datasets for different types of UAVs. Rozantsev et al. [8] proposed convolutional neural network-based regressors on spatiotemporal image cubes extracted using a multi-scale sliding window to detect flying objects. They demonstrated the robust performance of changes in the appearance of flying objects with low resolution, but the entire procedure could not be executed in real-time.
Thus, this paper proposes an online learning-based hybrid tracking method that combines detection and tracking to overcome the above challenges, and enables trackers to be more discriminative for flying object tracking. Object tracking is pertinent for localizing an object of interest by exploiting the object correspondence between frames; however, distractors, e.g., similar looking objects and appearance changes, can cause the tracker to shift focus from the target to the background. The proposed tracking method addresses this issue by integrating object detection to determine the locations and scales of objects in an image. However, this demands perceiving detections that could potentially belong to distractors because they are unfavorable to differentiate the target object from other objects in the same category. Thus, the proposed method integrates both a detector and a tracker by predicting the probability of the object’s presence using an online learning classifier. The classifier attempts to mitigate the limitations of the tracker and detector, while exploiting the benefits of each. In addition, the classifier updates the features in an online manner rather than prior learning, which improves robustness against abrupt changes in an object’s appearance. As a result, the proposed hybrid tracking method exhibits powerful localization capability by handling both mistracking and misdetection.
In particular, the main contributions of this paper can be summarized as follows: (1) This paper introduces a robust drone-tracking approach that integrates detecting and tracking algorithms based on an online learning classifier. (2) This paper evaluates the proposed method for the two main goals, which are tracking and detection, on our own drone dataset and drone-vs-bird dataset. The proposed method outperforms existing trackers. (3) This paper performs an evaluation on common aerial video datasets, which are UAV123 and UAVL, to demonstrate the generalization.
The remainder of this paper is organized as follows. Section 2 introduces work related to object detection and tracking. Section 3 describes the proposed hybrid tracking method, including the online learning-based integration method. Section 4 discusses experiments conducted on drone-based datasets to evaluate the proposed method. Finally, the paper is concluded in Section 5.

3. Methodology

3.1. Proposed Framework

In this paper, an online learning-based hybrid tracking method that combines detection and tracking is proposed to track UAVs effectively. Object tracking localizes the target according to the object correspondence between frames; however, the tracking drift problem frequently results in losing the target object. In addition, object detection, which localizes and classifies objects, has limitations relative to distinguishing the target from other objects in the same class, and handling the diverse sizes and appearances of UAVs. As discussed in Section 1, tracking a flying object, e.g., UAVs, in outdoor environments is challenging. To recover tracking when the tracker loses the target object, the proposed hybrid tracking method integrates both a detector and tracker. Here, the key is to derive the output by predicting the probability of objectness, which decides the likelihood of an object, via the online learning classifier. A powerful feature that could serve for discriminating the object from others is required in achieving robust tracking; thus, the classifier learns the object’s features during tracking to accommodate changes in the object’s appearance. This improves the ability to discriminate misdetection and mistracking, and it realizes more robust performance.
Figure 1 shows the framework of the proposed hybrid tracking method. The proposed method includes three main components, i.e., the detector (Section 2.1), the tracker (Section 2.2), and the integrator, which is used to learn the object’s features, predict their confidences, and switch the detector and tracker. The tracking process begins after the framework is initialized when an object of interest is given. First, the detector and tracker each predict a bounding box in a given image. To realize an effective balance between accuracy and processing speed, the detector is employed to locate objects either at regular frame intervals or in cases where the tracker fails. The controller is used to send each on/off signal associated with the execution to the detector and tracker. For convenience, the output bounding boxes of the detector, tracker, and integrator are expressed as follows:
B i = x 1 B i , y 1 B i , x 2 B i , y 2 B i , ( i = D , T , a n d   O ) ,
where x 1 B i , y 1 B i and x 2 B i , y 2 B i indicate the top-left and bottom-right coordinates of the bounding box, respectively. With B D and B T given by the detector and tracker, respectively, the integrator derives the result B O from the parameters of the strong classifier passed by the updater. Here, if the detector and tracker do not predict the target or perform by the passed on/off signal from the controller, the input to the integrator can be either B D or B T . To draw the output from B D and B T in the integrator, the strong classifier predicts c o n f D and c o n f T , which are the confidences of B D and B T , respectively. The confidence score reflects the objectness score that determines the likelihood of an object. With the confidence score, the target decision module derives the output B O based on a given threshold, which attempts to circumvent obstacles by discerning misdetection and mistracking, as described in the following two scenarios. In the first scenario, if one of B D and B T , whichever has the larger confidence score, is above the threshold, the bounding box associated with this confidence becomes the output B O . The output is then passed to the updater for online training of the parameters of the strong classifier, and to the tracker for the subsequent tracking process. For training, the patch generation collects training samples based on the output B O as the positive sample, and the surrounding regions represent negative samples. The details of the updater are discussed in Section 3.2.
Figure 1. Framework of the proposed hybrid tracking method.
In the second scenario, if the larger confidence score, either B D or B T , is unsatisfied based on the threshold, the target decision module considers all bounding boxes as the object’s absence, and then discards all bounding boxes.
The proposed hybrid tracking method can deploy diverse detection and tracking algorithms depending on the specific goals and circumstances of the corresponding application. For example, to detect small objects, e.g., UAVs, EfficientDet [37] and FPN [15] can be employed to extract semantically strong features by fusing multi-scale features. YOLO [12] and SSD [38] can benefit from their efficient and fast approaches. For the tracking, Siamese-based trackers can yield effective and powerful performance; however, non-deep learning algorithms, e.g., MedianFlow [39] and the KCF [22], are suitable for real-time tracking on onboard computers with limited hardware resources.

3.2. Online Classifier for Integration

In the proposed method, the online learning-based AdaBoost classifier is employed to integrate the detector and tracker, thereby mitigating their drawbacks by predicting their corresponding confidences. The AdaBoost algorithm has been applied in a wide range of machine learning tasks and applications as an ensemble algorithm [40]. AdaBoost builds a strong classifier as a linear combination of weak classifiers. Based on an online boosting approach [41], a variety of computer vision applications have been researched, e.g., object detection [42] and visual object tracking [40,43]. The core concept of the online boosting approach is that it updates all weak classifiers using a single training sample compared to offline boosting, which updates a single weak classifier using all available training samples. Based on [41], Grabner and Bischof [40] proposed the online AdaBoost method for feature selection using a strong classifier comprising selectors rather than a weak classifier. Here, each selector holds a weak classifier corresponding to the global feature pool. The primary purpose of the selector is exploiting online boosting to the selectors, not directly to the weak classifiers. They demonstrated that this technique realizes efficient computation of features during tracking, and allows the tracker to be performed in real-time. Inspired by [40,43], the proposed method implements an integrator with the online AdaBoost classifier to reduce the negative impact of the distractors.
The proposed hybrid tracking method classifies the bounding boxes from the detector and tracker by exploiting the parameters of a strong classifier passed by the updater, which is trained using samples generated by the previous tracking result. The strong classifier’s confidence is measured by the linear combination of selectors as follows:
c o n f x = n = 1 N α n · h n s e l x
where h n s e l and α n are n -th selector and its voting weight, respectively, given a set of N selectors, and x is the input patch by B D and B T . Training the updater means that weak classifiers are updated, and each selector chooses the best weak classifier with the lowest estimated error.
In the patch generation module (Figure 1), the training samples X = B 1 , B 2 , , B L are built for online adaptation of the classifier during tracking by exploiting B O as positive sample B 1 , and extracting a set of L 1 negative samples surrounding B O . Here, the l -th negative sample B l   l = 2,3 , L is defined as follows:
B l = x 1 B o + α 1 B l y 1 B o + β 1 B l x 2 B o + α 2 B l y 2 B o + β 2 B l
where x 1 B o , y 1 B o , x 2 B o , y 2 B o represent the coordinates of the bounding box of B O , and α 1 B l , β 1 B l , α 2 B l , β 2 B l represent the offsets of B l .
The algorithm for the training selectors’ procedure, which constitutes the strong classifier within the updater, is presented in Algorithm 1. The procedure is founded on online learning-based Adaboost [40,43]. For further details regarding the symbols and notations used in Algorithm 1, please refer to [40,43]. Note that the procedure of the training selectors updates the weak classifiers, selectors, and voting weights for classifiers. First, the importance weight λ of a training sample is initialized. Given training sample B l and its corresponding label y l 1,1 (a negative and positive sample, respectively), the set of M weak classifiers of the global feature pool is updated (lines 2–4). Here, each weak classifier classifies the sample and then updates its probability distributions for positive labeled samples and negative labeled samples, depending on the estimated label.
The selectors are updated as in lines 5–29. Given the responses of the weak classifiers h w e a k B l , the n -th selector h n s e l chooses the weak classifier h m + w e a k , which has the lowest error e m (lines 6–16). λ m c and λ m w represent the number of times that samples are classified correctly and incorrectly, respectively. The error is estimated from λ m c and λ m w . In accordance with the assumption of the boosting algorithm for binary classification, which stipulates that the error rate must be less than 50%, selectors exhibiting an error rate exceeding 50% are excluded from updates (line 17). Based on the error, the selector’ voting weight α n and the importance weight λ corresponding to the training sample are updated (lines 20–25). To adapt to changes in the object’s appearance and improve classification accuracy, the weak classifier with the highest error h m w e a k is initialized and replaced with a new randomly selected weak classifier. All selectors and corresponding weights are updated sequentially with the importance weight of the training sample (lines 26–28). As a result, the learned updater makes the classifier robust against object deformation between frames. In addition, the drift problems can be avoided by filtering out misdetections and mistracking. Here, Haar-like features [43] are used to generate weak hypotheses. Haar-like features can be obtained at low computational complexity using integral images as data structures, which helps realize real-time target classification during tracking [44].
 Algorithm 1: Online adapting classifier
 Require: training sample , B l , y l , y l { 1 , + 1 }
 1:    Set λ = 1
 2:    for  m = 1,2 , , M do
 3:        h m w e a k = u p d a t e ( h m w e a k , B l , y l )
 4:    end for
 5:    for  n = 1,2 , , N  do
 6:      for  m = 1,2 , , M do
 7:        if  h m w e a k B l = y l  then
 8:           λ m c = λ m c + λ
 9:        else
10:           λ m w = λ m w + λ
11:        end if
12:         e m = λ m w λ m c + λ m w
13:      end for
14:       m + = a r g m i n m ( e m )
15:       e n = e m +
16:       h n s e l = h m + w e a k
17:      if  e n = 0 or e n > 1 2  then
18:        exit
19:      end if
20:       α n = 1 2 · ln ( 1 e n e n )
21:      if  h m w e a k B l = y l  then
22:         λ = λ · 1 2 · ( 1 e n )
23:      else
24:         λ = λ · 1 2 · e n
25:      end if
26:    end for
27:     m = a r g m a x m ( e m )
28:     λ m c = λ m w = 1
29:    get new h m w e a k

4. Experiments and Results

4.1. Database

Most research based on supervised learning requires vast datasets with labeled ground truth data [45]. However, there is a lack of available datasets for UAVs, unlike more general objects, e.g., humans and vehicles. Thus, in this study, we constructed a unique dataset to track UAVs from outdoor environments. In addition, we also used the public drone-vs-bird dataset [46,47] to improve the proposed generalizability.
To construct our UAV dataset, we recorded videos of UAVs flying. Here, to ensure data diversity, the videos were captured according to three conditions (Figure 2), i.e., diverse backgrounds (cloud, buildings, mountain, etc.), drone types (Figure 3), and camera types (handheld, pan-tilt-zoom camera, etc.). The constructed UAV dataset contains 360 sequences with spatial resolution greater than or equal to full high definition (FHD). The dataset reflects many real-world challenges, including occlusion (OC), out-of-view (OV), background clutter (BC), camera motion (CM), low resolution (LR), scale variation (SV), and fast motion (FM), as described in Table 1. The dataset was created as part of a funded project, but there is a licensing issue with it. Unfortunately, the dataset is not public.
Figure 2. Screenshots taken from our dataset: (a) cloud background, (b) river background, (c) mountain background, (d) building background, (e) large target, (f) medium target, (g) small target, and (h) tiny target.
Figure 3. Examples of UAVs used to capture our dataset: (a) large UAV, (b) normal UAV, and (c) tiny UAV.
Table 1. Attribute annotation in dataset.
We used an image dataset to train the detector model. This image dataset was taken from our constructed dataset and was refined by obtaining a drone image per second to reduce overfitting, improve accuracy, and improve generalizability. This dataset, which was used as a training set and a testing set, includes 44,986 images from our constructed dataset, and 94,874 images from the drone-vs-bird dataset.
Table 2 shows a dataset for testing the tracking performance. This dataset comprises six videos (10,232 frames) from the drone-vs-bird dataset and six videos (18,278 frames) from our constructed dataset. Each video sequence is over 25 FPS and annotated with the seven attributes described in Table 1.
Table 2. Test dataset.

4.2. Experimental Environment

4.2.1. Evaluation Metrics

We evaluated the proposed method in terms of both tracking and detection. For the tracking evaluation, we used metrics from the Object Tracking Benchmark (OTB) (both 2013 [48] and 2015 [3] versions) and the Anti-UAV Benchmark [49]. The detection performance was compared using the following metrics from the PASCAL VOC [50] and ImageNet challenge [51].
  • IoU: The IoU is a measure of relative overlap between two bounding boxes. For example, if a tracked bounding box r t and ground truth bounding box r o of a target object are given, their IoU is defined as follows:
IoU r t , r o = r t r o r t r o
  • Center location error (CLE): The CLE is the Euclidean distance between a tracked center location and a manually labeled ground truth position.
For quantitative analysis, the most common evaluation metrics are precision and the success rate. In the precision plot, a frame is marked as being tracked successfully if the CLE score is less than a given threshold [48]. The success plot marks a successful frame if the IoU score is greater than a given threshold. The precision and success plots give the ratio of successful frames at the specified thresholds, and each plot is delineated by varying the threshold values [4]. In addition, the precision and success plots generally rank the tracking methods at thresholds equal to 20 and 0.5, respectively [48].
In real-world scenarios, there is an increased risk that the tracker drifts to the background caused by primary problems, such as occlusion and out-of-view [4]. However, when a tracking algorithm loses track of the target object, the output can be random, which causes incorrect evaluation of tracking performance [3]. To address these issues, the mean state accuracy ( m S A ) [49] was used in our evaluations. The state accuracy S A is measured for a given sequence as follows:
SA = t T I O U t × δ v t > 0 + p t × ( 1 δ v t > 0 ) T ,
where I o U t and v t are the I o U and ground truth visibility flag at frame t , respectively. If the target exists in frame t , δ v t > 0 = 1 ; otherwise, δ v t > 0 = 0 . If the tracker predicts that the target is absent, the p t value will be 1; otherwise, the p t value will be 0. The m S A is taken as the average S A value for all video sequences.
Recall that the proposed method employs a detector; thus, it can also be used for the detection task without requiring any adaptation at test time. Here, to evaluate detection performance, we used the F-measure, which is the harmonic mean of precision and recall [52]. Based on the true positive (TP), false positive (FP), and false negative (FN), precision, i.e., the percentage of correct positive predictions, indicates a model’s ability to recognize only relevant objects. Recall that the percentage of correct positive predictions among all given ground truths represents the model’s ability to discern all relevant cases [53]. The measurements classify a detection as correct or incorrect by comparing the I o U . In this study, the I o U threshold was set to 0.3. Details of the metrics are introduced in [24,52,53].

4.2.2. Implementation Details

In these experiments, the proposed method is implemented employing YOLOv4 [20] and MedianFlow [39] as the detector and tracker, respectively, to demonstrate performance that is suitable for surveillance systems, which commonly use low-resource onboard computers. MedianFlow executes at high speed on a single CPU, and YOLOv4 runs in real-time on a single GPU while obtaining high accuracy. It should be noted that the proposed hybrid framework is not restricted to the utilization of a particular detector and tracker. In the proposed method, the online learning-based strong classifier consists of 50 selectors, each with a feature pool of 10 weak classifiers. Images with 614 × 614 are fed to YOLOv4. We train YOLOv4 from scratch for 100 epochs with a batch size of 24 using the image dataset described in Section 4.1. Here, the dataset is randomly composed of 109,360 bounding boxes as the training set, and 30,500 bounding boxes as the testing set.

4.3. Experimental Comparison

In order to evaluate the performance of the proposed integration approach, we conducted a comparative analysis of its tracking and detection capabilities with those of MedianFlow [39], YOLOv4 [20], SiamRPN++ [30] (a deep learning-based tracking model renowned for its superior tracking accuracy), and a linear combination-based integration method [5], which we henceforth refer to as LC. Specifically, we compared the proposed approach against YOLOv4 and MedianFlow, owing to their widespread use in the literature. Additionally, we evaluated the performance of our method against SiamRPN++, which serves as a representative example of deep learning-based tracking models. Finally, we included LC in our analysis given its similarity in approach to our proposed method. This comparative analysis also demonstrates the effectiveness of our approach in enhancing the individual performance of commonly used detector and tracker components.

4.3.1. Evaluation for Object Tracking

To report the tracking algorithms’ performance in precision and success plots by varying the thresholds, the one-pass evaluation (OPE) technique was employed in this study.
The OPE technique runs an algorithm with initialization from the ground truth in the first frame and reports the average precision or success rate of all results [3,47]. Following the evaluation protocol of OPE [3], Figure 4 shows that the proposed method with MedianFlow and YOLOv4 achieved the best success score of 0.747 at a threshold value of 0.3. Thus, the proposed method outperformed the LC [5] (0.719) and YOLOv4 (0.618) methods. In addition, the proposed method produced the best precision score of 0.797 at a threshold value of 20. Compared to the LC method, the proposed method realized a 2.8% improvement in terms of overlap, and a 2.9% improvement in terms of precision. YOLOv4 obtained the best performance when the threshold in the success plot was greater than approximately 0.6, and when the threshold in the precision plot was less than 5. These results indicate that YOLOv4 most accurately predicted the bounding box. Russakovsky et al. reported that humans have difficulty distinguishing an IoU of 0.3 from 0.5 [54]. Accordingly, if the IoU threshold value of 0.3 is considered sufficient to realize steady tracking, tracking methods that incorporate both detection and tracking, e.g., the proposed method and LC method, will exhibit robust performance. In particular, the proposed method is more robust than the LC method, which only uses the location of the bounding box.
Figure 4. (a) Success plots of test sequences; (b) precision plots of test sequences.
As mentioned previously, a tracker can easily drift to the background due to real-world challenges, e.g., occlusion and out-of-view problems; thus, it is important to determine whether the object is, in fact, present. To evaluate this, we used the mSA metric. Table 3 shows that the proposed method achieved the best mSA score of 0.553, which indicates that the proposed method belongs to the top three in terms of SA for all test sequences. The proposed method was able to perceive the object’s absence. Even though the SiamRPN++ method obtained the best SA score for Seq3, Seq7, and Seq11, it is limited in terms of handling occlusions and out-of-view situations, as demonstrated by Seq1, Seq2, Seq4, Seq6, and Seq8. Occlusion and fully out-of-view cases are generally major problems in object tracking tasks. If the tracker does not employ a strategy to update the model, the tracker may easily drift from the correct target to various distractors. To prove that the proposed method can handle such distractors appropriately, the representative quantitative results of the proposed method are compared to those of SiamRPN++ in terms of IoU overlap in Figure 5. In Seq2, the object left the view and became occluded. The proposed and SiamRPN++ methods encounter the problem of drifting to the background after frame 195. As a result, the SiamRPN++ lost the target. However, the proposed method could recover the track in frame 225 by the detection mechanism. Figure 5 shows that the proposed method maintains the tracking after the occurrence of occlusion in frame 435, by taking advantage of the redetection process.
Table 3. Performance SA (%) and mSA (%) on test sequences (first-, second-, and third-ranked trackers are labeled in red, blue, and bold, respectively).
Figure 5. Frame-by-frame comparison of overlap score on Seq2. Yellow and purple shaded boxes indicate OV and OC, respectively. In the bottom, the red, light blue, and green boxes denote the proposed method (MedianFlow, YOLOv4), SiamRPN++ [30], and ground truth, respectively.
Table 4 compares the processing time per frame in FPS of the proposed approach to that of the MedianFlow, YOLOv4, and LC methods, in order to evaluate its computational complexity. The reported speed was obtained using an Intel i7-6700 CPU, and the YOLOv4 result was obtained using the Intel i7-6850K CPU and RTX 2080 SUPER GPU. In this experiment, the YOLOv4 is performed independently; hence, the speed of the proposed and LC methods was measured, except for the performance of YOLOv4. MedianFlow, which is a light model that achieves tracking by estimating the displacement of a number of points within the bounding box, ran at 40.84 FPS. Even though the speed of the LC method decreases by 84% of MedianFlow’s, it still ran at an effective real-time speed of 34.35 FPS. YOLOv4 achieved the best result of 49.46 FPS; however, this method suffers a critical limitation in terms of distinguishing the target object from other objects in the same class. As shown in Table 4, the proposed method obtained a comparably inferior speed of 21.90 FPS; however, the proposed method can run in effective real-time on applications that require greater than 20 FPS. In addition, the proposed method has demonstrated outstanding robustness; thus, it could offer a reasonable trade-off between accuracy and speed by adjusting the number of weak classifiers of boosting and the detector execution period.
Table 4. Comparison of FPS of methods on constructed dataset.

4.3.2. Evaluation for Object Detection

The proposed method was primarily designed for object tracking; however, it can also be applied to object detection. In reference to the literature [50,51], we tested the proposed method compared to the YOLOv4 and LC methods in terms of precision, recall, and F-measure.
As shown in Table 5, the proposed method outperformed the compared methods in terms of TP, recall, and F-measure. YOLOv4 predicts bounding boxes by exploiting a pretrained model on vast datasets. In this evaluation, YOLOv4 achieved the best precision value (0.930) derived from the lowest FP. The proposed and LC methods can suffer from drifting, as demonstrated by the FP results; however, by using a tracking algorithm, the proposed and LC methods exhibit dominant performance in terms of FN by exploiting the benefits of both the tracker and detector. Thus, the proposed method achieved high precision (0.836) following YOLOv4, and the best recall value. In other words, the proposed method can maintain steady tracking track while not losing the target. Relative to the precision and recall results, the proposed method achieved significant F-measure performance (0.879), which proves that the detector can improve accuracy by incorporating an object tracking algorithm.
Table 5. Comparison of precision, recall, and F-measure on test sequences.

4.3.3. Visual Comparison

A tracker must be able to handle various distractors to realize robust tracking in real-world applications. Figure 6 shows a visual comparison of the results obtained by the proposed method and other methods, i.e., MedianFlow, SiamRPN++, YOLOv4, and LC, on the test dataset. Note that all methods were initialized in the first frame for this evaluation.
Figure 6. Visual results of the compared methods (from top to bottom) for Seq4, Seq5, Seq6, and Seq8. Red boxes represent the proposed method (MedianFlow and YOLOv4), green boxes represent the ground truth, yellow boxes represent MedianFlow [39], pink boxes represent YOLOv4 [20], light-blue boxes represent SiamRPN++ [30], and blue boxes represent the LC (MedianFlow, YOLOv4) method [5].
In the first row, Seq4 demonstrates that MedianFlow and SiamRPN++ can fail to track the target in out-of-view cases. Here, as the target disappears, the trackers (except MedianFlow) identify the target’s absence, but MedianFlow drifts to the background in frame 1319. Then, when the target is revealed, the proposed, LC, and YOLOv4 methods can locate the target accurately with more reasonable localization confidence due to the use of the detection mechanism. These results demonstrate that using only a single tracker, e.g., SiamRPN++, cannot realize long-term visual tracking in out-of-view cases without employing a redetection mechanism.
In Seq5, the proposed, LC, and SiamRPN++ methods quickly adapted to the fast motion and great scale variations of the target, despite the low resolution, while MedianFlow and YOLOv4 failed to realize the same adaptation. This result proves that the YOLOv4 method suffers when handling small objects.
Seq6 illustrates the behavior of the methods in the out-of-view case caused by rapid camera movement. After frame 162, MedianFlow estimated the oversized bounding box compared with the ground truth. In this case, SiamRPN++’s target drifted to the background; however, it recovered the tracking after the target appeared in the search area. Note that the proposed and LC methods recognized the target’s absence and maintained tracking from the position at which the YOLOv4 method detected the target’s reappearance.
Finally, the results obtained for Seq7 demonstrate that the target became occluded by a building structure and then moved behind it. In addition, several distractors of similar flying objects frequently appeared in this case. As shown in Figure 6, from the beginning, MedianFlow gets confused while tracking the tiny target (here, the target was only approximately 80 pixels). YOLOv4 failed to detect the target by predicting a another similar object as a drone; hence, the LC method failed to track because it is incapable of recognizing the changes in the target’s appearance. In contrast, the proposed method was able to locate the target accurately by handling incorrect detection results, using the integration mechanism that exploits the online learned classifier.
Thus, we believe that the proposed method demonstrates beneficial use of the online updating mechanism and the combined detection and tracking method framework.

4.4. Evaluation on Other Datasets

We evaluate the proposed method on datasets commonly used for aerial tracking, such as UAV123 [55] and UAVL [55]. The UAV123 and UAVL datasets include low-altitude aerial videos captured from a UAV. The datasets reflect scenarios of common visual tracking challenges, such as aspect ratio change (ARC), background clutter (BC), fast motion (FM), full occlusion (FOC), partial occlusion (POC), illumination variation (IV), low resolution (LR), out-of-view (OV), similar objects (SOB), scale variation (SV), and viewpoint change (VC). These datasets are employed to assess whether a tracker is suitable in real-world scenarios [34].
As same as the evaluation in Section 4.3, we compare the proposed method with YOLOv4, MedianFlow, and the aforementioned integration method called LC, in terms of success and precision plots. The integration methods of the proposed method and LC employ YOLOv4 and MedianFlow as the detector and tracker, respectively. In the UAV123 and UAVL experiments, the model of YOLOv4 was trained with the COCO dataset [56]. In accordance with classes on the COCO dataset, we use 77 sequences from the UAV123 dataset and 14 sequences that include cars, trucks, boats, bikes, persons, birds, and aerial vehicles classes, except building and group classes.

4.4.1. Results on UAV123 Dataset

Figure 7 shows the success and precision plots on the UAV123 dataset. The proposed method shows the best success score and second-best precision score. Figure 8 shows the success plots of the proposed and other methods for the 12 scenarios. The proposed method achieves the best success scores in 8 out of 12 scenarios, including SV, ARC, POC, CM, FM, LR, IV, and SOB. The proposed method dominantly outperforms in FM, LR, IV, and SOB scenarios.
Figure 7. (a) Success plots on UAV123 dataset; (b) precision plots on UAV123 dataset.
Figure 8. Success plots on different scenarios in UAV123 dataset.
The reason for the high success score under SV, ARC, IV, and POC includes that the target’s appearance changes are treated well by the proposed online learning. Our online learned classifier also allows handling SOB. As shown in FM and CM, the proposed method’s detector helps to re-track when the target is missing. Our tracking under LR, thanks to the MedianFlow, is suitable for tiny object tracking. This comparison shows that the proposed method takes advantage of the detector and the tracker. On the other hand, the performance of the proposed method is slightly less than the YOLOv4 or LC method in FOC, OV, BC, and VC, likely because model updating is often accompanied by a dilemma. Note that a model may not catch the appearance changes without online updating, but the updating by an appearance that is not the intended target could bring the risk of encountering the problem of drifting.

4.4.2. Results on UAVL Dataset

Figure 9 and Figure 10 illustrate the experimental results on the UAVL dataset. The proposed method achieves promising results in 10 out of 12 scenarios, including FOC, POC, IV, VC, CM, SV, ARC, LR, FM, and SOB. Among other things, the proposed method shows outstanding results in SV, ARC, LR, FM, and SOB. They reveal that the proposed method also works well in long-term tracking, since it is able to track again with the help of the detector even if tracking drifts.
Figure 9. (a) Success plots on UAVL dataset; (b) precision plots on UAVL dataset.
Figure 10. Success plots on different scenarios in UAVL dataset.

4.5. Limitations

The proposed hybrid UAV tracking method, while effective in many scenarios, still has limitations. In scenes with a high density of dynamic and tiny objects, the detection and tracking methods may fail depending on their respective abilities. Wrong estimation of the detection and tracking methods may also contaminate the online-learned classifier. Once the classifier’s ability for discrimination is weakened, there is an increased risk of the proposed hybrid tracking method drifting to the background, whether semantic or not. Another real-world application concern is processing time, since the proposed framework requires detection and tracking mechanisms. Therefore, implementing multi-threading could help reduce time consumption.

5. Conclusions

This paper has proposed a hybrid UAV tracking method that combines detection and tracking techniques to handle common distractors, by exploiting an online classifier-based integration method. In the proposed method, the classifier updates features in an online manner; thus, the method is robust against object deformations, e.g., appearance changes and scale variations.
To evaluate the proposed method, we constructed a custom dataset containing images of UAVs flying in outdoor environments. In addition, we considered a public drone-vs-bird dataset to train and test the tracking methods. The experimental results demonstrate that the proposed method is efficient and accurate. The proposed hybrid method with the YOLOv4 and MedianFlow techniques has demonstrated the importance of the online updating mechanism and combined detection and tracking method framework. This framework can be applied to other computer vision tasks, e.g., object detection, depending on their goals. As the proposed framework can be applied with any detection and tracking techniques, the improvement of object detection and tracking is expected to improve our method. If the detection and tracking algorithms integrated in the proposed framework have a very limited performance, our method may also be subject to their performance restriction.

Author Contributions

Conceptualization and methodology, S.S. and H.C.; project administration, I.L. and J.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by a grant from Police Science and Technology R&D Program funded by the Korean National Police Agency [21PCRD-B128275-05, Development and field Demonstration Test of Surveillance System using radar and EO/IR for detecting illegal Flight of UAVs] and the Institute of Information and Communications Technology Planning and Evaluation (IITP) grant funded by the Korean government (MSIT) [2020-0-00452, Development of Adaptive Viewer-centric Point cloud AR/VR(AVPA) Streaming Platform].

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
UAVunmanned aerial vehicles
IoUintersection-over-union
RoIregion of interest
RPNregion proposal network
FPSframes per second
KCFkernelized correlation filter
SiamFCfully convolutional Siamese network
CFEcenter location error
mSAmean state accuracy
OPEone-pass evaluation
ARCaspect ratio change
BCbackground clutters
FMfast motion
FOCfull occlusion
POCpartial occlusion
IVillumination variation
LRlow resolution
OVout-of-view
SOBsimilar objects
SVscale variation
VCview-point change

References

  1. Kang, H.; Joung, J.; Kim, J.; Kang, J.; Cho, Y.S. Protect your sky: A survey of counter unmanned aerial vehicle systems. IEEE Access 2020, 8, 168671–168710. [Google Scholar] [CrossRef]
  2. Shin, J.M.; Kim, Y.S.; Ban, T.W.; Choi, S.; Kang, K.M.; Ryu, J.Y. Position tracking techniques using multiple receivers for anti-drone systems. Sensors 2020, 21, 35. [Google Scholar] [CrossRef] [PubMed]
  3. Wu, Y.; Lim, J.; Yang, M.H. Object tracking benchmark. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 37, 1834–1848. [Google Scholar] [CrossRef] [PubMed]
  4. Ondrašovič, M.; Tarábek, P. Siamese visual object tracking: A survey. IEEE Access 2021, 9, 110149–110172. [Google Scholar] [CrossRef]
  5. Wang, Q.; Rasmussen, C.; Song, C. Fast, deep detection and tracking of birds and nests. In Proceedings of the Advances in Visual Computing: 12th International Symposium, Las Vegas, NV, USA, 12–14 December 2016. [Google Scholar]
  6. Saribas, H.; Uzun, B.; Benligiray, B.; Eker, O.; Cevikalp, H. A hybrid method for tracking of objects by UAVs. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, Long Beach, CA, USA, 16–20 June 2019. [Google Scholar]
  7. Baptista, M.; Fernandes, L.; Chaves, P. Tracking and Classification of Aerial Objects. In Proceedings of the Intelligent Transport Systems, from Research and Development to the Market Uptake: Third EAI International Conference, INTSYS 2019, Braga, Portugal, 4–6 December 2019. [Google Scholar]
  8. Rozantsev, A.; Lepetit, V.; Fua, P. Detecting flying objects using a single moving camera. IEEE Trans. Pattern Anal. Mach. Intell. 2016, 39, 879–892. [Google Scholar] [CrossRef] [PubMed]
  9. Liu, L.; Ouyang, W.; Wang, X.; Fieguth, P.; Chen, J.; Liu, X.; Pietikäinen, M. Deep learning for generic object detection: A survey. Int. J. Comput. Vis. 2020, 128, 261–318. [Google Scholar] [CrossRef]
  10. Jiao, L.; Zhang, F.; Liu, F.; Yang, S.; Li, L.; Feng, Z.; Qu, R. A survey of deep learning-based object detection. IEEE Access 2019, 7, 128837–128868. [Google Scholar] [CrossRef]
  11. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster r-cnn: Towards real-time object detection with region proposal networks. Adv. Neural Inf. Process. Syst. 2015, 28, 91–99. [Google Scholar] [CrossRef] [PubMed]
  12. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You only look once: Unified, real-time object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 26 June–1 July 2016. [Google Scholar]
  13. Girshick, R.; Donahue, J.; Darrell, T.; Malik, J. Rich feature hierarchies for accurate object detection and semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 24–27 June 2014. [Google Scholar]
  14. Girshick, R. Fast r-cnn. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 13–16 December 2015. [Google Scholar]
  15. Lin, T.Y.; Dollár, P.; Girshick, R.; He, K.; Hariharan, B.; Belongie, S. Feature pyramid networks for object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017. [Google Scholar]
  16. Dai, J.; Li, Y.; He, K.; Sun, J. R-fcn: Object detection via region-based fully convolutional networks. Adv. Neural Inf. Process. Syst. 2016, 29, 379–387. [Google Scholar]
  17. He, K.; Gkioxari, G.; Dollár, P.; Girshick, R. Mask r-cnn. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017. [Google Scholar]
  18. Redmon, J.; Farhadi, A. YOLO9000: Better, faster, stronger. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017. [Google Scholar]
  19. Redmon, J.; Farhadi, A. Yolov3: An incremental improvement. arXiv 2018, arXiv:1804.02767. [Google Scholar]
  20. Bochkovskiy, A.; Wang, C.Y.; Liao, H.Y.M. Yolov4: Optimal speed and accuracy of object detection. arXiv 2020, arXiv:2004.10934. [Google Scholar]
  21. Aker, C.; Kalkan, S. Using deep networks for drone detection. In Proceedings of the 2017 14th IEEE International Conference on Advanced Video and Signal Based Surveillance (AVSS), Lecce, Italy, 29 August–1 September 2017. [Google Scholar]
  22. Henriques, J.F.; Caseiro, R.; Martins, P.; Batista, J. High-speed tracking with kernelized correlation filters. IEEE Trans. Pattern Anal. Mach. Intell. 2014, 37, 583–596. [Google Scholar] [CrossRef] [PubMed]
  23. Grabner, H.; Grabner, M.; Bischof, H. Real-time tracking via on-line boosting. In Proceedings of the Bmvc, Edinburgh, Scotland, 4–7 September 2006. [Google Scholar]
  24. Marvasti-Zadeh, S.M.; Cheng, L.; Ghanei-Yakhdan, H.; Kasaei, S. Deep learning for visual tracking: A comprehensive survey. IEEE Trans. Intell. Transp. Syst. 2021, 23, 3943–3968. [Google Scholar] [CrossRef]
  25. Bertinetto, L.; Valmadre, J.; Henriques, J.F.; Vedaldi, A.; Torr, P.H. Fully-convolutional siamese networks for object tracking. In Proceedings of the Computer Vision–ECCV 2016 Workshops, Amsterdam, The Netherlands, 8–10, 15–16 October 2016. [Google Scholar]
  26. Li, Y.; Zhang, X. SiamVGG: Visual tracking using deeper siamese networks. arXiv 2019, arXiv:1902.02804. [Google Scholar]
  27. Guo, Q.; Feng, W.; Zhou, C.; Huang, R.; Wan, L.; Wang, S. Learning dynamic siamese network for visual object tracking. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017. [Google Scholar]
  28. He, A.; Luo, C.; Tian, X.; Zeng, W. A twofold siamese network for real-time object tracking. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018. [Google Scholar]
  29. Li, B.; Yan, J.; Wu, W.; Zhu, Z.; Hu, X. High performance visual tracking with siamese region proposal network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018. [Google Scholar]
  30. Li, B.; Wu, W.; Wang, Q.; Zhang, F.; Xing, J.; Yan, J. Siamrpn++: Evolution of siamese visual tracking with very deep networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–20 June 2019. [Google Scholar]
  31. Zhu, Z.; Wang, Q.; Li, B.; Wu, W.; Yan, J.; Hu, W. Distractor-aware siamese networks for visual object tracking. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018. [Google Scholar]
  32. Guo, D.; Wang, J.; Cui, Y.; Wang, Z.; Chen, S. SiamCAR: Siamese fully convolutional classification and regression for visual tracking. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Virtual Conference, 14–19 June 2020. [Google Scholar]
  33. Wang, Q.; Zhang, L.; Bertinetto, L.; Hu, W.; Torr, P.H. Fast online object tracking and segmentation: A unifying approach. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–20 June 2019. [Google Scholar]
  34. Chen, Z.; Zhong, B.; Li, G.; Zhang, S.; Ji, R. Siamese box adaptive network for visual tracking. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Virtual Conference, 14–19 June 2020. [Google Scholar]
  35. Sosnovik, I.; Moskalev, A.; Smeulders, A.W. Scale equivariance improves siamese tracking. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Virtual Conference, 5–9 January 2021. [Google Scholar]
  36. Çintaş, E.; Özyer, B.; Şimşek, E. Vision-based moving UAV tracking by another UAV on low-cost hardware and a new ground control station. IEEE Accesss 2020, 8, 194601–194611. [Google Scholar] [CrossRef]
  37. Tan, M.; Pang, R.; Le, Q.V. Efficientdet: Scalable and efficient object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Virtual Conference, 14–19 June 2020. [Google Scholar]
  38. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.Y.; Berg, A.C. Ssd: Single shot multibox detector. In Proceedings of the Computer Vision–ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, 11–14 October 2016. [Google Scholar]
  39. Kalal, Z.; Mikolajczyk, K.; Matas, J. Forward-backward error: Automatic detection of tracking failures. In Proceedings of the 2010 20th International Conference on Pattern Recognition, Istanbul, Turkey, 23–26 August 2010. [Google Scholar]
  40. Grabner, H.; Bischof, H. On-line boosting and vision. In Proceedings of the 2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, New York, NY, USA, 17–22 June 2006. [Google Scholar]
  41. Oza, N.C.; Russell, S.J. Online bagging and boosting. In Proceedings of the International Workshop on Artificial Intelligence and Statistics, Key West, FL, USA, 4–7 January 2001. [Google Scholar]
  42. Nguyen, T.T.; Grabner, H.; Bischof, H.; Gruber, B. On-line boosting for car detection from aerial images. In Proceedings of the 2007 IEEE International Conference on Research, Innovation and Vision for the Future, Hanoi, Vietnam, 5–9 March 2007. [Google Scholar]
  43. Viola, P.; Jones, M. Rapid object detection using a boosted cascade of simple features. In Proceedings of the 2001 IEEE Computer society Conference on Computer Vision and Pattern Recognition, Kauai, HI, USA, 8–14 December 2021. [Google Scholar]
  44. Porikli, F. Integral histogram: A fast way to extract histograms in cartesian spaces. In Proceedings of the 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, San Diego, CA, USA, 20–25 June 2005. [Google Scholar]
  45. Taha, B.; Shoufan, A. Machine learning-based drone detection and classification: State-of-the-art in research. IEEE Access 2019, 7, 138669–138682. [Google Scholar] [CrossRef]
  46. Coluccia, A.; Fascista, A.; Schumann, A.; Sommer, L.; Dimou, A.; Zarpalas, D.; Akyon, F.C.; Eryuksel, O.; Ozfuttu, K.A.; Altinuc, S.O.; et al. Drone-vs-bird detection challenge at IEEE AVSS2021. In Proceedings of the 2021 17th IEEE International Conference on Advanced Video and Signal Based Surveillance (AVSS), Virtual Conference, 16–19 November 2021. [Google Scholar]
  47. Drone-vs-Bird Detection Challenge. Available online: https://wosdetc2020.wordpress.com/drone-vs-bird-detectionchallenge/ (accessed on 15 March 2020).
  48. Wu, Y.; Lim, J.; Yang, M.H. Online object tracking: A benchmark. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Oregon, Portland, 23–28 June 2013. [Google Scholar]
  49. Jiang, N.; Wang, K.; Peng, X.; Yu, X.; Wang, Q.; Xing, J.; Li, G.; Zhao, J.; Guo, G.; Han, Z. Anti-UAV: A large multi-modal benchmark for UAV tracking. arXiv 2021, arXiv:2101.08466. [Google Scholar]
  50. Everingham, M.; Van Gool, L.; Williams, C.K.; Winn, J.; Zisserman, A. The pascal visual object classes (voc) challenge. Int. J. Comput. Vis. 2009, 88, 303–308. [Google Scholar] [CrossRef]
  51. Deng, J.; Dong, W.; Socher, R.; Li, L.J.; Li, K.; Li, F.-F. Imagenet: A large-scale hierarchical image database. In Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009. [Google Scholar]
  52. Powers, D.M. Evaluation: From precision, recall and F-measure to ROC, informedness, markedness and correlation. arXiv 2020, arXiv:2010.16061. [Google Scholar]
  53. Padilla, R.; Netto, S.L.; Da Silva, E.A. A survey on performance metrics for object-detection algorithms. In Proceedings of the 2020 International Conference on Systems, Signals and Image Processing (IWSSIP), Rio de Janeiro, Brazil, 1–3 July 2020. [Google Scholar]
  54. Russakovsky, O.; Li, L.J.; Li, F.-F. Best of both worlds: Human-machine collaboration for object annotation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015. [Google Scholar]
  55. Mueller, M.; Smith, N.; Ghanem, B. A benchmark and simulator for uav tracking. In Proceedings of the Computer Vision–ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, 11–14 October 2016. [Google Scholar]
  56. Lin, T.Y.; Maire, M.; Belongie, S.; Hays, J.; Perona, P.; Ramanan, D.; Bourdev, L.; Girshick, R.; Dollár, P.; Zitnick, C.L. Microsoft coco: Common objects in context. In Proceedings of the Computer Vision–ECCV 2014: 13th European Conference, Zurich, Switzerland, 6–12 September 2014. [Google Scholar]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.