Next Article in Journal / Special Issue
All-Purpose Nano- and Microcontainers: A Review of the New Engineering Possibilities
Previous Article in Journal / Special Issue
Dispersive Optical Solitons for Stochastic Fokas-Lenells Equation With Multiplicative White Noise
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Evaluation of Color Anomaly Detection in Multispectral Images for Synthetic Aperture Sensing

Institute of Computer Graphics, Johannes Kepler University Linz, 4040 Linz, Austria
*
Author to whom correspondence should be addressed.
Eng 2022, 3(4), 541-553; https://doi.org/10.3390/eng3040038
Submission received: 3 November 2022 / Revised: 24 November 2022 / Accepted: 25 November 2022 / Published: 29 November 2022
(This article belongs to the Special Issue Feature Papers in Eng 2022)

Abstract

:
In this article, we evaluate unsupervised anomaly detection methods in multispectral images obtained with a wavelength-independent synthetic aperture sensing technique called Airborne Optical Sectioning (AOS). With a focus on search and rescue missions that apply drones to locate missing or injured persons in dense forest and require real-time operation, we evaluate the runtime vs. quality of these methods. Furthermore, we show that color anomaly detection methods that normally operate in the visual range always benefit from an additional far infrared (thermal) channel. We also show that, even without additional thermal bands, the choice of color space in the visual range already has an impact on the detection results. Color spaces such as HSV and HLS have the potential to outperform the widely used RGB color space, especially when color anomaly detection is used for forest-like environments.

1. Introduction

Color anomaly detection methods identify pixel regions in multispectral images that have a low probability of occurring in the background landscape and are therefore considered to be outliers. Such techniques are used in remote sensing applications for agriculture, wildlife observation, surveillance, or search and rescue. Occlusion caused by vegetation, however, remains a major challenge.
Airborne Optical Sectioning (AOS) [1,2,3,4,5,6,7,8,9,10,11,12,13] is a synthetic aperture sensing technique that computationally removes occlusion in real-time by registering and integrating multiple images captured within a large synthetic aperture area above the forest (cf. Figure 1). With the resulting shallow-depth-of-field integral images, it becomes possible to locate targets (e.g., people, animals, vehicles, wildfires, etc.) that are otherwise hidden under the forest canopy. Image pixels that correspond to the same target on the synthetic focal plane (i.e., the forest ground) are computationally aligned and enhanced, while occluders above the focal plane (i.e., trees) are suppressed in strong defocus. AOS is real-time and wavelength-independent (i.e., it can be applied to images in all spectral bands), which is beneficial for many areas of application. Thus far, AOS has been applied to the visible [1,11] and the far-infrared (thermal) spectrum [4] for various applications, such as archeology [1,2], wildlife observation [5], and search and rescue [8,9]. By employing a randomly distributed statistical model [3,10,12], the limits of AOS and its efficacy with respect to its optimal sampling parameters can be explained. Common image processing tasks, such as classification with deep neural networks [8,9] or color anomaly detection, [11] are proven to perform significantly better when applied to AOS integral images compared with conventional aerial images. We also demonstrated the real-time capability of AOS by deploying it on a fully autonomous and classification-driven adaptive search and rescue drone [9]. In [11,13], we presented the first solutions to tracking moving people through densely occluding foliage.
Anomaly detection methods for wilderness search and rescue have been evaluated earlier [14], and bimodal systems using a composition of visible and thermal information were already used to improve detection rates of machine learning algorithms [15,16]. However, none of the previous work considered occlusion.
With AOS, we are able to combine multispectral recordings into a single integral image. Our previous work has shown that image processing tasks, such as person classification with deep neural networks [8,9,10], perform significantly better on integral images when compared to single images. These classifiers are based on supervised architectures, which have the disadvantage that training data must be collected and labeled in a time-consuming manner and that the trained neural networks do not generalize well into other domains. It was also shown in [11] that the image integration process of AOS decreases variance and covariance, which allows better separation of target and background pixels when applying the Reed–Xiaoli (RX) unsupervised anomaly detection [17].
In this article, we evaluate several common unsupervised anomaly detection methods being applied to multispectral integral images that are captured from a drone when flying over open and occluded (forest) landscapes. We show that their performance can significantly be improved by the right combination of spectral bands and choice of color space input format. Especially for forest-like environments, detection rates of occluded people can be consistently increased if visible and thermal bands are combined and if HSV or HLS color spaces are used for the visible bands instead of common RGB. Furthermore, we also evaluate the runtime behavior of these methods when considered for time-critical applications, such as search and rescue.

2. Materials and Methods

For our evaluation, we applied the dataset from [8], which was used to prove that integral images improve people classification under occluded conditions. It consists of RGB and thermal images (pairwise simultaneously) captured with a drone prototype over multiple forest types (broadleaf, conifer, mixed) and open landscapes, as shown in Figure 2. In all images, targets (persons laying on the ground) are manually labeled. Additional telemetry data (GPS and IMU sensor values) of the drone during capturing are also provided for each image.
While the visible bands were converted from RGB to other color spaces (HLS, HSV, LAB, LUV, XYZ, and YUV), the thermal data were optionally added as a fourth (alpha) channel, resulting in additional input options (RGB-T, HLS-T, HSV-T, LAB-T, LUV-T, XYZ-T, and YUV-T).
All images had a resolution of 512x512 pixels, so the input dimensions where either (512, 512, 3) or (512, 512, 4). Methods that do not require spatial information used flattened images with (262144, 3) or (262144, 4) dimensions.
The publicly available C/C++ implementation of AOS Source Code: https://github.com/JKU-ICG/AOS (accessed on 24 November 2022) was used to compute integral images from single images.

2.1. Color Anomaly Detectors

Unsupervised color anomaly detectors have been widely used in the past [17,18,19,20,21,22], with the Reed–Xiaoli (RX) detector [17] being commonly considered as a benchmark. Several variations of RX exist, where the standard implementation calculates global background statistics (over the entire image) and then compares individual pixels based on the Mahalanobis distance. In the further course of this article, we will refer to this particular RX detector as Reed–Xiaoli Global (RXG).
The following briefly summarizes the considered color anomaly detectors, while details can be found through the provided references:
The Reed–Xiaoli Global (RXG) detector [17] computes a K n × n covariance matrix of the image, where n is the number of input channels (e.g., for RGB, n = 3 and for RGB-T, n = 4 ). The pixel under test is the n-dimensional vector r, and the mean is given by the n-dimensional vector μ :
α R X G ( r ) = ( r μ ) T K n × n 1 ( r μ ) .
The Reed–Xiaoli Modified (RXM) detector [18] is a variation of RXG, where an additional constant κ = | | r μ | | 1 is used for normalization:
α R X M ( r ) = κ · α R X G ( r ) = r μ | | r μ | | T K n × n 1 ( r μ ) .
The Reed–Xiaoli Local (RXL) detector computes covariance and mean over smaller local areas and, therefore, does not use global background statistics. The areas are defined by an inner window ( g u a r d _ w i n ) and an outer window ( b g _ w i n ). The mean μ and covariance K are calculated based on the outer window but excludes the inner window. Window sizes were chosen to be g u a r d _ w i n = 33 and b g _ w i n = 55 , based on the projected pixel sizes of the targets in the forest landscape.
The principal component analysis (PCA) [19] uses singular value decomposition for a linear dimensionality reduction. The covariance matrix of the image is decomposed into eigenvectors and their corresponding eigenvalues. A low-dimensional hyperplane is constructed by selected ( n _ c o m p o n e n t s ) eigenvectors. Outlier scores for each sample are then obtained by their euclidean distance to the constructed hyperplane. The number of eigenvectors to use was chosen to be n _ c o m p o n e n t s = n , where n is the number of input channels (e.g., for RGB, n = 3 and for RGB-T, n = 4 ).
The Gaussian mixture model (GMM) [20] is a clustering approach, where multiple Gaussian distributions are used to characterize the data. The data are fit to each of the single Gaussians ( n _ c o m p o n e n t s ), which are considered as a representation of clusters. For each sample, the algorithm calculates the probability of belonging to each cluster, where low probabilities are an indication of being an anomaly. The number of Gaussians to use was chosen to be n _ c o m p o n e n t s = 2 .
The cluster based anomaly detection (CBAD) [21] estimates background statistics over clusters instead of sliding windows. The image background is partitioned (using any clustering algorithm) into clusters ( n _ c l u s t e r ), where each cluster can be modeled as a Gaussian distribution. Similar to GMM, anomalies have values that deviate significantly from the cluster distributions. Samples are each assigned to the nearest background cluster, becoming an anomaly if their value deviates farther from the mean than background pixel values in that cluster. The number of clusters to use was chosen to be n _ c l u s t e r = 2 .
The local outlier factor (LOF) [22] uses a distance metric (e.g., Minkowski distance) to determine the distances between neighboring ( n _ n e i g h b o r s ) data points. Based on the inverse of those average distances, the local density is calculated. This is then compared to the local densities of their surrounding neighborhood. Samples that have significantly lower densities than their neighbors are considered isolated and, therefore, become outliers. The number of neighbors to use was chosen to be n _ n e i g h b o r s = 200 .

2.2. Evaluation

The evaluation of the methods summarized above was carried out on a consumer PC (Intel Core i9-11900H @ 2.50GHz) for the landscapes shown in Figure 2.
Precision (Equation (1)) vs. recall (Equation (2)) was used as metrics for performance comparisons.
The task can be formulated as a binary classification problem, where positive predictions are considered anomalous pixels. The data we want to classify (image pixels) is highly unbalanced, as most of the pixels are considered as background (majority class) and only some of the pixels are considered anomalies (minority class).
The true positive (TP) pixels are determined by checking whether they lie within one of the labeled rectangles, as shown in Figure 2. Pixels detected outside these rectangles are considered false positives (FP) and pixels inside the rectangle but not classified anomalously are considered false negatives (FN). Since the dataset only provides rectangles for labels and not perfect masks around the persons, the recall results are biased (in general, not as good as expected). As we are mainly interested in the performance difference between individual methods and the errors introduced are always constant (rectangle area—real person mask), the conclusions drawn from the results should be the same, even if perfect masks were used instead.
The precision (Equation (1)) quantifies the number of correct positive predictions made, and recall (Equation (2)) quantifies the number of correct positive predictions made out of all positive predictions that could have been made.
P r e c i s i o n = T P T P + F P ,
R e c a l l = T P F N + T P .
Precision and recall both focus on the minority class (anomalous pixels) and are therefore less concerned with the majority class (background pixels), which is important for our unbalanced dataset.
Since the anomaly detection methods provide probabilistic scores on the likelihood of a pixel being considered anomalous, a threshold value must be chosen to obtain a final binary result.
The precision–recall curve (PRC) in Figure 3 shows the relationship between precision and recall for every possible threshold value that could be chosen. Thus, a method performing well would have high precision and high recall over different threshold values. We use the area under the precision–recall curve (AUPRC), which is simply the integral of the PRC, as the final evaluation metric.
The AUPRC metric provides comparable results on the overall performance of a method but is not well suited when it comes to finding the best threshold for a single image. To obtain the best threshold value for a single image, we use the F β -score (Equation (3)), which is also calculated from precision and recall:
F β = 1 + β 2 · P r e c i s i o n · R e c a l l β 2 · P r e c i s i o n + R e c a l l ,
where β is used as a weighting factor that can be chosen such that recall is considered β -times more important than precision.
The balanced F1-score is the harmonic mean of precision and recall and is widely used. However, as we care more about minimizing false positives than minimizing false negatives, we would select a β < 1 . A grid search of β values has given the best results for β = 1 2 . With this setting, precision is weighted more heavily than recall. The F β metric is only used to threshold the image scores for comparison purposes, as shown in Figure 6.

3. Results

Figure 4 shows the AUPRC values across different color spaces and methods. The methods are evaluated on each color space, once with three channels (visible spectrum only) and once with four channels (visible and thermal spectrum). The results of the forest landscape are average values over F0, F1 and F5, and the results of the open landscape are average values over O1, O3 and O5.
As expected (and as we have also seen in Figure 3), the overall AUPRC of the open landscapes is much higher than the AUPRC values of the more challenging forest landscapes. The reason is an occlusion in the presence of forests.
The AUPRC values of the four-channel (color + thermal) and three-channel (color only) inputs are overlayed in the same bar. The slightly lighter colored four-channel results are always higher than the three-channel results—regardless of the method or the color space used. However, the difference is more pronounced for the forest landscapes than for the open landscapes. This shows that regardless of the scenery and regardless of the method and the color space used, the additional thermal information always improves the performance of anomaly detection.
With a look at the AUPRC values in the forest landscapes, we can observe that RXL gives the overall best results and outperforms all other methods. Utilizing the additional thermal information gives, in this case, even a 2 × gain. This can also be observed visually in the anomaly detection scores shown in Figure 6, where FP’s detections highly decrease and TP’s detections highly increase if the thermal channel is added (e.g., F1, in the visible spectral band, many background pixels are considered anomalous, with the additional thermal information those misclassified pixels are eliminated).
Looking at the AUPRC values in the open landscapes, we can observe that the difference between the methods is not as pronounced as in the forest landscapes. An obvious outlier, however, seems to be LOF, which nevertheless performs very well (second best) in the forest landscapes. This can be explained by the fact that hyper-parameters of the methods were specifically chosen for the forest landscape. In the case of LOF, the n _ n e i g h b o r s parameter was set to be 200, which seems suboptimal for the open landscapes. The same holds for RXL (window sizes), CBAD (number of clusters), GMM (number of components) and PCA (number of components). All other methods do not require hyper-parametrization.
Another observation that can be made is that some color spaces consistently give better results than others. In the forest landscapes, HSV(-T) usually gives the best results, regardless of the methods being used. In the open landscapes, it is not as clear which color space performs best, but HSV(-T) still gives overall good results. In general, and especially for RXM, the improvements achieved by choosing HSV(-T) over other color spaces are clearly noticeable.
The individual results plotted in Figure 4 are also shown in Table 1 and Table 2, where the mean values over all color spaces (last row) may give a useful estimate of the method’s overall performance. The highest AUPRC value for the forest and open scenery is highlighted in bold.
Since anomaly detection for time-critical applications should deliver reliable results in real-time, we have also measured their runtimes, as shown in Table 3. The best-performing methods on the forest landscapes in terms of AUPRC values are RXL and LOF. In terms of runtime, both are found to be very slow, as they consume 20 to 35 s for computations, where all other algorithms provide anomaly scores in under a second (cf. Figure 5).

4. Discussion

The AUPRC results in Figure 4 show that all color anomaly detection methods benefit from additional thermal information, but especially in combination with the forest landscapes.
In challenging environments, where the distribution of colors has a much higher variance (e.g., F1 in Figure 6, due to bright sunlight), the additional thermal information improves results significantly. If the temperature difference between targets and the surrounding is large enough, the thermal spectral band may add spatial information (e.g., distinct clusters of persons), which is beneficial for methods that calculate results based on locality properties (e.g., RXL, LOF).
In forest-like environments, the RXL anomaly detector performs best regardless of the input color space. This could be explained by the specific characteristics of an integral image. In the case of occlusion, the integration process produces highly blurred images caused by defocused occluders (forest canopy) above the ground, which results in a much more uniformly distributed background. Since target pixels on the ground stay in focus, anomaly detection methods such as RXL, which calculate background statistics on a smaller window around the target, benefit from the uniform distributed (local) background. The same is true for LOF, where the local density in the blurred background regions is much higher than the local density in the focused target region, resulting in overall better outlier detection rates. Since most objects in open landscapes are located near the focal plane (i.e., at nearly the same altitude above the ground), there is no out-of-focus effect caused by the integration process. Thus, these methods do not produce similarly good results for open landscapes.
For the forest landscapes, the HSV(-T) and HSL(-T) color spaces consistently give better results than others. The color spaces HSV (hue, saturation, value) and HSL (hue, saturation, lightness) are both based on cylindrical color space geometries and differ mainly in their last dimension (brightness/lightness). The first two dimensions (hue, saturation) can be considered more important when distinguishing colors, as the last dimension only describes the value (brightness) or lightness of a color. We assume that the more uniform background resulting from the integration process also has a positive effect on the distance metric calculations when those two color spaces are used, especially if the background mainly consists of a very similar color tone. This is again more pronounced for the forest landscapes than for the open landscapes.
Although the AUPRC results obtained from RXL and LOF are best for forest landscapes, the high runtime indicates that these methods are impractical for real-time applications. A trade-off must be made between good anomaly detection results and fast runtime; therefore, we consider the top-performing methods that provide reliable results within milliseconds further.
Based on the AUPRC and runtime results shown in Figure 5, one could suggest that the RXM method may be used. The AUPRC results combined with HSV-T are the best among methods that run under one second, regardless of the landscape. Since this method does not require a-priory settings to be chosen (only the final thresholding value) and the runtime is one of the fastest, it would be well suited for usage in forests and open landscapes. The second-best algorithm based on the AUPRC values would be CBAD, with the disadvantage that it requires a hyper-parameter setting and does not generalize well for open landscapes.

5. Conclusions

In this article, we have shown that the performance of unsupervised color anomaly detection methods applied to multispectral integral images can be further improved by an additional thermal channel. Each of the evaluated methods performs significantly better when thermal information is utilized in addition, regardless of the landscape (forest or open). Another finding is that even without the additional thermal band, the choice of input color space (for the visible channels) already has an influence on the results. Color spaces such as HSV and HLS can outperform the widely used RGB color space, especially in forest-like landscapes.
These findings might guard decisions on the choice of color anomaly detection method, input format, and applied spectral band, depending on individual use cases. Occlusion cause by vegetation, such as forests, remains challenging for many of them. In the future, we will investigate anomalies caused by motion in the context of synthetic aperture sensing. In combination with color and thermal anomaly detection, motion anomaly detection has the potential to further improve detection results for moving targets, such as people, animals, or vehicles.

Author Contributions

Conceptualization, O.B. and F.S.; methodology, F.S.; software, F.S. and I.K.; validation, F.S., I.K. and O.B.; formal analysis, F.S.; investigation, F.S.; resources, I.K.; data curation, I.K.; writing—original draft preparation, F.S. and O.B.; writing—review and editing, F.S. and O.B.; visualization, F.S.; supervision, O.B.; project administration, O.B.; funding acquisition, O.B. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Austrian Science Fund (FWF) under grant number P32185-NBL and by the State of Upper Austria and the Austrian Federal Ministry of Education, Science and Research via the LIT–Linz Institute of Technology under grant number LIT-2019-8-SEE114.

Data Availability Statement

The data and source code used in the experiments can be downloaded from https://doi.org/10.5281/zenodo.3894773 and https://github.com/JKU-ICG/AOS (accessed on 24 November 2022).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Kurmi, I.; Schedl, D.; Bimber, O. Airborne optical sectioning. J. Imaging 2018, 4, 102. [Google Scholar] [CrossRef] [Green Version]
  2. Bimber, O.; Kurmi, I.; Schedl, D. Synthetic aperture imaging with drones. IEEE Comput. Graph. Appl. 2019, 39, 8–15. [Google Scholar] [CrossRef] [PubMed]
  3. Kurmi, I.; Schedl, D.; Bimber, O. A statistical view on synthetic aperture imaging for occlusion removal. IEEE Sens. J 2019, 19, 9374–9383. [Google Scholar] [CrossRef] [Green Version]
  4. Kurmi, I.; Schedl, D.; Bimber, O. Thermal airborne optical sectioning. Remote Sens. 2019, 11, 1668. [Google Scholar] [CrossRef] [Green Version]
  5. Schedl, D.; Kurmi, I.; Bimber, O. Airborne optical sectioning for nesting observation. Sci. Rep. 2020, 10, 7254. [Google Scholar] [CrossRef]
  6. Kurmi, I.; Schedl, D.; Bimber, O. Fast Automatic Visibility Optimization for Thermal Synthetic Aperture Visualization. IEEE Geosci. Remote Sens. Lett. 2021, 18, 836–840. [Google Scholar] [CrossRef]
  7. Kurmi, I.; Schedl, D.; Bimber, O. Pose Error Reduction for Focus Enhancement in Thermal Synthetic Aperture Visualization. IEEE Geosci. Remote. Sens. Lett. 2021; to be published. [Google Scholar] [CrossRef]
  8. Schedl, D.; Kurmi, I.; Bimber, O. Search and rescue with airborne optical sectioning. Nat. Mach. Intell. 2020, 2, 783–790. [Google Scholar] [CrossRef]
  9. Schedl, D.; Kurmi, I.; Bimber, O. An autonomous drone for search and rescue in forests using airborne optical sectioning. Sci. Robot 2021, 6, 1188. [Google Scholar] [CrossRef]
  10. Kurmi, I.; Schedl, D.; Bimber, O. Combined person classification with airborne optical sectioning. Sci. Rep. 2022, 12, 3804. [Google Scholar] [CrossRef]
  11. Nathan, R.; Kurmi, I.; Schedl, D.; Bimber, O. Through-Foliage Tracking with Airborne Optical Sectioning. J. Remote Sens. 2022, 2022, 9812765. [Google Scholar] [CrossRef]
  12. Seits, F.; Kurmi, I.; Nathan, R.; Ortner, R.; Bimber, O. On the Role of Field of View for Occlusion Removal with Airborne Optical Sectioning. arXiv 2022, arXiv:2204.13371. [Google Scholar] [CrossRef]
  13. Nathan, R.; Kurmi, I.; Bimber, O. Inverse Airborne Optical Sectioning. Drones 2022, 6, 231. [Google Scholar] [CrossRef]
  14. Morse, B.; Thornton, D.; Goodrich, M. Color anomaly detection and suggestion for wilderness search and rescue. In Proceedings of the Seventh Annual ACM/IEEE International Conference on Human-Robot Interaction, Boston, MA, USA, 5–8 March 2012; Association for Computing Machinery: New York, NY, USA, 2012; pp. 455–462. [Google Scholar] [CrossRef] [Green Version]
  15. Rudol, P.; Doherty, P. Human Body Detection and Geolocalization for UAV Search and Rescue Missions Using Color and Thermal Imagery. In Proceedings of the 2008 IEEE Aerospace Conference, Big Sky, MT, USA, 1–8 March 2008; pp. 1–8. [Google Scholar] [CrossRef]
  16. Hinzmann, T.; Stegemann, T.; Cadena, C.; Siegwart, R. Deep Learning-based Human Detection for UAVs with Optical and Infrared Cameras: System and Experiments. arXiv 2020, arXiv:2008.04197. [Google Scholar] [CrossRef]
  17. Reed, I.; Yu, X. Adaptive Multiple-Band CFAR Detection of an Optical Pattern with Unknown Spectral Distribution. IEEE Trans. Acoust. Speech Signal Process. 1990, 38, 1760–1770. [Google Scholar] [CrossRef]
  18. Chang, C.; Chiang, S. Anomaly detection and classification for hyperspectral imagery. Geosci. Remote Sens. IEEE Trans. 2002, 40, 1314–1325. [Google Scholar] [CrossRef] [Green Version]
  19. Shyu, M.; Chen, S.; Sarinnapakorn, K.; Chang, L. A Novel Anomaly Detection Scheme Based on Principal Component Classifier. In Proceedings of the IEEE Foundations and New Directions of Data Mining Workshop, in conjunction with the Third IEEE International Conference on Data Mining (ICDM’03), Melbourne, FL, USA, 19–22 November 2003. [Google Scholar]
  20. Bishop, C.M.; Nasrabadi, N.M. Pattern Recognition and Machine Learning. J. Electron. Imaging 2007, 16, 049901. [Google Scholar]
  21. Carlotto, M. A cluster-based approach for detecting man-made objects and changes in imagery. IEEE Trans. Geosci. Remote Sens. 2005, 43, 374–387. [Google Scholar] [CrossRef]
  22. Breunig, M.; Kriegel, H.; Ng, R.; Sander, J. LOF: Identifying Density-Based Local Outliers. ACM SIGMOD Rec. 2000, 29, 93–104. [Google Scholar] [CrossRef]
Figure 1. Airborne optical sectioning (AOS) is a synthetic aperture sensing technique that computationally combines multiple aerial images captured within a synthetic aperture area (a) to an integral image, which enhances targets on the synthetic focal plane while suppressing occluders above it. Right: People covered by forest canopy. Single aerial image (thermal channel) that suffers from strong occlusion (b), and corresponding integral image of the same environment with occlusion removed (c).
Figure 1. Airborne optical sectioning (AOS) is a synthetic aperture sensing technique that computationally combines multiple aerial images captured within a synthetic aperture area (a) to an integral image, which enhances targets on the synthetic focal plane while suppressing occluders above it. Right: People covered by forest canopy. Single aerial image (thermal channel) that suffers from strong occlusion (b), and corresponding integral image of the same environment with occlusion removed (c).
Eng 03 00038 g001
Figure 2. Our evaluation dataset consists of several forest (a) and open (b) landscape images captured with a drone from an altitude of about 35m AGL. Each scenery (F0, F1, F5, O1, O3, O5) contains about 20 consecutive single images taken in the visible (RGB) and thermal spectrum, which are combined into integral images. Rectangles indicate manually labeled persons lying on the ground.
Figure 2. Our evaluation dataset consists of several forest (a) and open (b) landscape images captured with a drone from an altitude of about 35m AGL. Each scenery (F0, F1, F5, O1, O3, O5) contains about 20 consecutive single images taken in the visible (RGB) and thermal spectrum, which are combined into integral images. Rectangles indicate manually labeled persons lying on the ground.
Eng 03 00038 g002
Figure 3. The area under the precision–recall curve (AUPRC) is used as a metric for comparing the performance of the evaluated anomaly detection methods. We consider true positives (TP), false positives (FP), and false negatives (FN) pixels for each image and calculate precision and recall. The example above illustrates precision–recall curves of all landscapes for RXG and with RGB-T input.
Figure 3. The area under the precision–recall curve (AUPRC) is used as a metric for comparing the performance of the evaluated anomaly detection methods. We consider true positives (TP), false positives (FP), and false negatives (FN) pixels for each image and calculate precision and recall. The example above illustrates precision–recall curves of all landscapes for RXG and with RGB-T input.
Eng 03 00038 g003
Figure 4. Results of area under the precision–recall curve (AUPRC) values for multiple color spaces and color anomaly detection methods. The results of the forest landscape are average values over F0, F1, and F5, and the results of the open landscape are average values over O1, O3, and O5. The stacked bar charts highlight the improvement gains caused by the additional thermal channel.
Figure 4. Results of area under the precision–recall curve (AUPRC) values for multiple color spaces and color anomaly detection methods. The results of the forest landscape are average values over F0, F1, and F5, and the results of the open landscape are average values over O1, O3, and O5. The stacked bar charts highlight the improvement gains caused by the additional thermal channel.
Eng 03 00038 g004
Figure 5. Performance in AUPRC (left bars) vs. runtime in m s (right bars): Color anomaly detection methods that produce results in less than a second. Reed–Xiaoli local (RXL) and local outlier factor (LOF) performed well but needed more than 20 seconds and are, therefore, not practicable for applications with real-time demands.
Figure 5. Performance in AUPRC (left bars) vs. runtime in m s (right bars): Color anomaly detection methods that produce results in less than a second. Reed–Xiaoli local (RXL) and local outlier factor (LOF) performed well but needed more than 20 seconds and are, therefore, not practicable for applications with real-time demands.
Eng 03 00038 g005
Figure 6. Color anomaly detection scores for forest (a) and open (b) landscapes, comparing the overall good performing HSV(-T) inputs. The first rows per scenery show anomaly scores of the best.
Figure 6. Color anomaly detection scores for forest (a) and open (b) landscapes, comparing the overall good performing HSV(-T) inputs. The first rows per scenery show anomaly scores of the best.
Eng 03 00038 g006
Table 1. Area under the precision–recall curve (AUPRC) values for each color space and color anomaly detection method. The scores are obtained from integral images and are averaged over forest landscapes. The last row is the mean AUPRC value over all color spaces. The best performance is highlighted in bold.
Table 1. Area under the precision–recall curve (AUPRC) values for each color space and color anomaly detection method. The scores are obtained from integral images and are averaged over forest landscapes. The last row is the mean AUPRC value over all color spaces. The best performance is highlighted in bold.
CBADGMMLOFPCARXGRXLRXM
HLS0.0470.0290.0370.0280.0290.0930.044
HSV0.0500.0330.0400.0280.0330.0960.048
LAB0.0290.0280.0380.0230.0280.0770.031
LUV0.0320.0320.0380.0240.0320.0980.034
RGB0.0290.0270.0360.0260.0270.0840.030
XYZ0.0270.0270.0390.0270.0270.0880.027
YUV0.0290.0260.0380.0230.0260.0770.032
AUPRC0.0350.0290.0380.0250.0290.0880.035
CBADGMMLOFPCARXGRXLRXM
HLS-T0.0810.0610.0980.0430.0620.2180.099
HSV-T0.0930.0760.0990.0530.0770.2230.108
LAB-T0.0830.0640.0840.0520.0650.1960.047
LUV-T0.0760.0560.0830.0470.0560.1860.052
RGB-T0.0840.0570.0920.0320.0570.1910.055
XYZ-T0.0820.0570.0830.0320.0580.1960.037
YUV-T0.0780.0550.0880.0420.0560.1900.050
AUPRC0.0820.0610.0900.0430.0620.2000.064
Table 2. Area under the precision–recall curve (AUPRC) values for each color space and anomaly detection method. The scores are obtained from integral images and are averaged over open landscapes. The last row is the mean AUPRC value over all color spaces. Best performance is highlighted in bold.
Table 2. Area under the precision–recall curve (AUPRC) values for each color space and anomaly detection method. The scores are obtained from integral images and are averaged over open landscapes. The last row is the mean AUPRC value over all color spaces. Best performance is highlighted in bold.
CBADGMMLOFPCARXGRXLRXM
HLS0.2190.2550.0220.2390.2560.2150.275
HSV0.2110.2560.0440.2460.2580.2170.291
LAB0.2430.2420.0210.2760.2410.1440.189
LUV0.2340.2300.0200.2290.2280.1360.185
RGB0.2610.2570.0780.0940.2560.1530.175
XYZ0.2540.2540.0190.0780.2530.1550.144
YUV0.2570.2560.0190.2490.2550.1530.162
AUPRC0.2400.2500.0320.2020.2490.1680.203
CBADGMMLOFPCARXGRXLRXM
HLS-T0.2580.2930.0770.2920.2940.2680.296
HSV-T0.2590.3030.0820.3020.3040.2680.316
LAB-T0.3050.3090.0820.3060.3100.2220.250
LUV-T0.2950.3020.0820.2670.3000.2120.249
RGB-T0.3120.3120.0860.1870.3120.2250.260
XYZ-T0.3080.3100.1230.1690.3100.2290.219
YUV-T0.3140.3120.0750.2970.3120.2250.250
AUPRC0.2930.3060.0870.2600.3060.2360.263
Table 3. Runtime for each input format and method in milliseconds. The input format (color spaces) does not have an influence on the runtime, but addition channels (thermal) may increase the runtime for some algorithms. The last row is the mean runtime of an algorithm. Best performance is highlighted in bold.
Table 3. Runtime for each input format and method in milliseconds. The input format (color spaces) does not have an influence on the runtime, but addition channels (thermal) may increase the runtime for some algorithms. The last row is the mean runtime of an algorithm. Best performance is highlighted in bold.
CBADGMMLOFPCARXGRXLRXM
HLS88721918,440983936,64728
HSV88322518,0941024036,34627
LAB87721919,833993636,49529
LUV88821219,975964136,16629
RGB92521619,367964035,73229
XYZ87222419,1071004036,36727
YUV87422820,4851004236,40927
Runtime88722119,329994036,30928
CBADGMMLOFPCARXGRXLRXM
HLS-T87824328,0541044336,00230
HSV-T88223728,2351074035,80130
LAB-T87923028,0051083235,87631
LUV-T89322927,5061054536,05932
RGB-T90423027,323983735,61832
XYZ-T88024124,3951084336,04729
YUV-T87723827,1461074036,18029
Runtime88523627,2381054035,94031
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Seits, F.; Kurmi, I.; Bimber, O. Evaluation of Color Anomaly Detection in Multispectral Images for Synthetic Aperture Sensing. Eng 2022, 3, 541-553. https://doi.org/10.3390/eng3040038

AMA Style

Seits F, Kurmi I, Bimber O. Evaluation of Color Anomaly Detection in Multispectral Images for Synthetic Aperture Sensing. Eng. 2022; 3(4):541-553. https://doi.org/10.3390/eng3040038

Chicago/Turabian Style

Seits, Francis, Indrajit Kurmi, and Oliver Bimber. 2022. "Evaluation of Color Anomaly Detection in Multispectral Images for Synthetic Aperture Sensing" Eng 3, no. 4: 541-553. https://doi.org/10.3390/eng3040038

APA Style

Seits, F., Kurmi, I., & Bimber, O. (2022). Evaluation of Color Anomaly Detection in Multispectral Images for Synthetic Aperture Sensing. Eng, 3(4), 541-553. https://doi.org/10.3390/eng3040038

Article Metrics

Back to TopTop