Next Article in Journal
A Comparison of Dense and Sparse Optical Flow Techniques for Low-Resolution Aerial Thermal Imagery
Previous Article in Journal
Fuzzy Information Discrimination Measures and Their Application to Low Dimensional Embedding Construction in the UMAP Algorithm
Previous Article in Special Issue
Seamless Copy–Move Replication in Digital Images
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Human Tracking in Top-View Fisheye Images: Analysis of Familiar Similarity Measures via HOG and against Various Color Spaces

1
EuroMov Digital Health in Motion, Univ Montpellier, IMT Mines Ales, 30100 Ales, France
2
College of Arts and Sciences, University of Nizwa, Nizwa 616, Oman
*
Author to whom correspondence should be addressed.
J. Imaging 2022, 8(4), 115; https://doi.org/10.3390/jimaging8040115
Submission received: 21 January 2022 / Revised: 16 March 2022 / Accepted: 29 March 2022 / Published: 16 April 2022
(This article belongs to the Special Issue Edge Detection Evaluation)

Abstract

:
The purpose of this paper is to find the best way to track human subjects in fisheye images by considering the most common similarity measures in the function of various color spaces as well as the HOG. To this end, we have relied on videos taken by a fisheye camera wherein multiple human subjects were recorded walking simultaneously, in random directions. Using an existing deep-learning method for the detection of persons in fisheye images, bounding boxes are extracted each containing information related to a single person. Consequently, each bounding box can be described by color features, usually color histograms; with the HOG relying on object shapes and contours. These descriptors do not inform the same features and they need to be evaluated in the context of tracking in top-view fisheye images. With this in perspective, a distance is computed to compare similarities between the detected bounding boxes of two consecutive frames. To do so, we are proposing a rate function ( S ) in order to compare and evaluate together the six different color spaces and six distances, and with the HOG. This function links inter-distance (i.e., the distance between the images of the same person throughout the frames of the video) with intra-distance (i.e., the distance between images of different people throughout the frames). It enables ascertaining a given feature descriptor (color or HOG) mapped to a corresponding similarity function and hence deciding the most reliable one to compute the similarity or the difference between two segmented persons. All these comparisons lead to some interesting results, as explained in the later part of the article.

1. Introduction

In the computer vision domain, detecting and tracking people constitutes an important area [1]. Significant approaches exist to detect pedestrians in monocular images [2]. Recently, numerous papers have been published on people or pedestrian detection using deep learning techniques. In this paper, our aim is to estimate the similarity between several small images containing persons by extracting color features. Indeed, color spaces have a “low computational” cost and must be investigated to choose the ideal space for a specific application [3,4,5,6]. The proposed study focuses on the extraction of people’s bounding boxes from top-view fisheye images. Usually, with its specific lens (see Figure 1), a fisheye camera offers panoramic views of 2 π radian angles [7,8]. Therefore, objectives with wide-angle lenses capture images that are typically warped, as illustrated in Figure 2. Besides the deformations, a challenging task is to tackle the fact that images may differ significantly due to variations in factors like illumination and human posture. Moreover, in the top view, with the camera axis pointing vertically, people standing straight may appear oriented in the image, pointing towards the image’s center due to the distortion of the camera lens. Therefore, a conventional perspective human detection technique such as the histogram of oriented gradients (HOG) cannot be directly used [9]. Various methods [10,11,12,13,14] have been implemented in the literature for distorted perspectives but they do not enable the tracking of the movement of a person throughout the video acquired by a fisheye camera, especially in a top view perspective.
In this paper, a comparison of the HOG and different usual color spaces with different similarity functions for histograms is presented in the context of human segmentation tracking via a fisheye camera. This communication is the continuation of our previous paper [15], in which only two color spaces are compared with only one similarity function. Our main objective is to determine the best combination of the similarity functions and color spaces that will enable better tracking of segmented persons via color features. The detection part is performed via a deep learning algorithm presented in [12]. Finally, experiments are carried out by OpenCV software and a function S enables an objective comparison of the results.
The rest of the paper is organized as follows. The next section presents the different color spaces. Then Section 3 is devoted to detailing the HOG. The similarity functions are introduced in Section 4. Finally, Section 5 presents the experimental protocol and different results before conclude this paper in Section 6.

2. Colorimetric Spaces

Standard color images are represented with red, green, and blue channels (RGB). Typically, a digital color image has a minimum of 24 bits/pixel, with 8 bits assigned to each color component of the three-color channels. Consequently, a color image can contain 65,536 different color shades. RGB is a standard color space to represent images on displays, cameras, projectors, etc. Although RGB is the primary color space used to describe the spectral content of color signals, a variety of other representations have been developed, as detailed in [16]. Most of them are summarized in [17]. Their use is focused on different image classification areas: face detection, food quality (fruits, vegetables, wine, honey...), medical images, and scene understanding are important to note, among others.
Some of these representations are more appropriate for image processing than others. The objective of this paper is to optimize color space/distance pair for better pedestrian tracking with fisheye cameras. We will limit this preliminary study to six well-known color spaces outlined in Table 1 (also see Appendix A) with special reference to the involvement of the HOG. The following subsections describe four types of these color spaces.

2.1. The Additive Models: RGB

As introduced above, the mixture of the three primary colors (red, green, and blue) allows the production of all the color shades perceptible to the human eye. Considering a color image representation where each channel is coded with 8 bits, the RGB space offers a model in which each pixel is represented by a triplet (R, G, B) with the intensity of each pixel component ranging from 0 to 255. Note that OpenCV reads images in BGR format (instead of RGB) but the treatments are the same as in the presented study.

2.2. The Natural Models: HSV and HLS

The RGB space can be visualized as a color cube, considering the basis is formed by the red, green, and blue vectors. Cylindrical color models that remap the RGB primary colors into dimensions are easy to understand visually. Indeed, HSV and HSL models are called natural because they are closer to the human perception of colors. In these models each pixel is coded by a triplet (H, S, V) or (H, L, S); for both spaces, the value H represents the hue. Hue is defined as the dominant wavelength of a source. Consider a “color wheel” around which we concatenate all the colors, so each primary color would be spaced 120 , the hue is represented by an angle (between 0 and 360 ), tied to the color of the pixel. As images are coded in 8 bits, so we will take for H half of its real value (so that its value is between 0 and 255). The S value, also common to both spaces, represents the saturation; it expresses the color contrast. It is between 0 and 255 (0 corresponds to gray whatever the color hue, and 255 a point on the color wheel). The last element of the triplet corresponds to the lightness and value (V or S): the higher it is, the clearer the image is.

2.3. The Human Vision Models: L*a*b* and L*u*v*

Many mixtures in the RGB space give indistinguishable colors. The International Commission on Illumination (CIE) has introduced three-dimensional representations, (including L*a*b* and L*u*v*, which we will study) that take the human perception for reference. Indeed, for a point of these spaces, L corresponds to the lightness, whereas a and b (or u and v) represent the chroma coordinates. L*a*b* and L*u*v* spaces are quite similar, a and b (or u and v) coordinates measure positions along with the green/red and blue/yellow axes. These spaces are difficult to understand intuitively. Since we use images coded in 8 bits, the values of each element of the triplet coding each pixel are reduced between 0 and 255.

2.4. The Television Models: YCbCr/YUV

The YCbCr model is one of the models used for digital or analog compression. The Y value represents the luminance that describes the image in black and white and the values (Cr, Cb) code the chrominance. Cr describes the red–green difference and Cb the blue–yellow difference. The Y component is more sensitive to the human eye, it must remain precise to distinguish colors correctly. However, Cb and Cr are less sensitive to the human eye, so they do not need to be more accurate. That is why this space is useful in image compression; especially in the JPEG algorithm.
Table 1 summarizes the different acronyms of the utilized color spaces, along with their ranges for OpenCV implementation. The appendix at the bottom recalls the way to compute each color space from RGB. Further, Figure 2 shows a sub-image of a segmented person under different color spaces.
Having described all the types of color spaces, we will now introduce the other feature descriptor that we will be comparing, namely the HOG.

3. Histogram of Oriented Gradients (HOG)

Contemporary machine learning hinges largely on what is now widely known as feature engineering, which simply means to derive newer features, from those in hand, for the refinement of the underlying model. With a considerable amount of data involved in multimedia, especially images and videos, the available raw features need to be ‘compressed’ to a simplified representation called the feature descriptors. In object detection, an important descriptor is the HOG, which relies on the shape and the contours of the object. Unlike simple edge descriptors, HOG is dependent on the orientation, in addition to the magnitude of the gradient. In practice, the image or the frame could be partitioned into regions, and HOG is computed for each region separately. As explained in [9], a typical edge detection via gradient for a grey level image involves the following steps:
  • Smooth the input image ( f ( x , y ) ), e.g., via a Gaussian function ( G ( x , y ) ) to get f ^ ( x , y ) :
    f ^ ( x , y ) = f ( x , y ) G ( x , y ) .
  • Compute the x and y derivatives of f ^ ( x , y )
    f ^ x ( x , y ) = f ( x , y ) D x ( x , y ) .
    f ^ y ( x , y ) = f ( x , y ) D y ( x , y ) .
    where D x and D x are the kernels for first order derivative in x and y directions, respectively. The familiar ones are Robert, Sobel, and Prewitt, 1 0 1 masks, etc. Such masks enables a computation of precise locations of edges of small objects and their tied orientations (see evaluation of edge detectors in [18]).
  • The magnitude of gradient m a g n ( x , y ) is given by:
    m a g n ( x , y ) = f ^ x 2 ( x , y ) + f ^ y 2 ( x , y ) .
    A simplified version is however preferred, which is:
    m a g n ( x , y ) = | f ^ x ( x , y ) | + | f ^ y ( x , y ) | .
  • For each pixel, the direction of gradient θ is computed as:
    θ ( x , y ) = tan 1 f ^ y ( x , y ) f ^ x ( x , y ) .
The resultant magnitude ( m a g n ( x , y ) ) and orientation ( θ ) maps can be used to compute HOG descriptors.
A simple approach to rely only on the orientation map and realize the histogram is to make a frequency table based on the orientation angles, either individually or in bins. However, the magnitude map can also be brought into play, if one replaces the frequency with the magnitude, against each angle (or a bin thereof). In such a case, a bin approach might be more feasible, but a given magnitude may have two adjacent candidate bins; it is better to assign to the nearest neighbor or divide among the bins based on nearness.
Usually, each image map will be partitioned to small fix-sized dyadic cells ( 8 × 8 or 16 × 16 or more) and HOG descriptors would be computed for each cell from its orientation map (in correspondence to its magnitude map, if needed). In essence, each of the cells has its own histogram and HOG descriptor. Better normalize the HOG descriptors, for better results, by dividing each bin value by the square root of the sum of the squares of all the bin values for the cell.
The HOG descriptor is usually a vector ( k × 1 matrix) of the form [ b 1 , b 2 , . , b k ] , where k is the number of bins and b i denotes the value in ith bin. A normalized vector for a given cell could be:
b 1 p , b 2 p , . , b k p ,
where
p = b 1 2 + b 2 2 + . b k 2 .
An m × n image, if divided into w × w blocks would have a total of m / w × n / w = m n / w 2 blocks. Since each block has a vector of k × 1 features, we end up with k × m n / w 2 features in total.
The HOG is useful to recognize people and objects in images [9]; it could be also combined with different strategies for the tracking process [19]. Consequently, the aim of this paper is to investigate if the HOG enables more precise histogram-based tracking of persons using a top-view moving fisheye camera than color spaces. In this study, we chose to implement the HOG on full gray-scale images, we used the BGR2GRAY OpenCV function to convert color images into scalar images. Considering this function, for each pixel tied to the (B, G, R) component of a color image, the pixel Y corresponding to the gray-scale image is computed such that:
Y = 0.299 · R + 0.587 · G + 0.114 · B .
greenNote that the filters we used to implement the HOG are [ 1 0 1 ] and [ 1 0 1 ] t and the gradient magnitude is computed with Equation (4); greensee the whole description in Figure 3.
After the description of the feature descriptors studied in this paper, now let us look at similarity functions.

4. Similarity Functions: Our Approach to Compare Histograms

This work is based on the comparison of HOG and six different color spaces, as a function of each of the chosen six different distance measures. To this end, the input is a video from a fisheye camera that is captured under the assumption that a limited number (2–4) of people are moving under it. In brief, the following steps were involved:
(i)
Employ the RAPiD method [12] to detect the different people in the frames of the video in the form of the bounding boxes.
(ii)
Convert each bounding box from RGB space to each of the desired spaces.
(iii)
In a given color space, compute the histograms H 1 and H 2 corresponding to the two histograms tied to two different images. They both contain N bins, and for an integer k between 0 and N 1 , H i ( k ) represents the value of the k-th bin (for i { 1 , 2 } )
(iv)
Normalize the histograms (i.e., k = 0 N 1 H i ( k ) = 1 ) in order to enable an easy comparison of two images, even of different sizes.
(v)
Compute the average of the histogram ( H ¯ i ) which is easily computed (optional, see Table 2):
H ¯ i = 1 N · k = 0 N 1 H i ( k ) .
(vi)
With H 1 , H 2 , H ¯ 1 , and H ¯ 2 as inputs, compute the identified six similarity measures (detailed bellow and in Table 2) for each of the chosen six color spaces.
(vii)
Compute the identified six similarity measures for HOG.
(viii)
Carry out a grid based comparison between the distance measures and color spaces as well as HOG using S function to decide the optimum.
Our comparison is based on six different similarity measures summarized in Table 2. These measures are in fact mostly distances and it will not be out of place to explain these briefly.
  • Histogram Correlation: The correlation measure, called correlation coefficient, is often used in statistics to determine if there is any interdependence between two random variables. It is defined as the quotient between their covariance and the product of their standard deviation. Originally, the result is between −1 and 1, where 1 indicates a perfect match and −1 a complete mismatch. To standardize with other distances, in particular, the complement of this measure is computed so that it translates to a value close to 0, for better matching.
  • Chi-square ( χ 2 ) measure: The chi-square ( χ 2 ) test statistic is a commonly used statistical measure to calculate the similarity between frequency distributions. The original chi-square measure, based on the Pearsons’s chi-square test statistic, is not only asymmetrical but also raises a singularity problem if either or both of the two histograms in the comparison contain a zero bit. That is why it was modified to nullify both the problems. This alternative version gives a measure of 0 for a perfect match and on the contrary, gives a score close to 1 for two very distant histograms (or images).
  • Histogram Intersection: The straightforward histogram intersection measure gives the proportion of pixels that have the same intensity. Therefore, a perfect match is obviously given by a score of 1. Like the correlation, the complement of this measure is being computed in this work.
  • Bhattacharyya distance: This measure estimates the similarity between two probability distributions. It has a geometric interpretation: considering two n-dimensional unit vectors p ( i ) and q ( i ) , this distance can be interpreted as a cosine of the angle between vectors formed by taking the square root of each coordinate of the initial vectors i.e., the angle formed by p ( i ) and q ( i ) . This measure is regularly used for object tracking [20].
  • Kullback–Leibler divergence: The Kullback–Leibler (KL) divergence is a measure of dissimilarity between two probability distributions. Considering two images, KL can be interpreted as a divergence by the amount of information lost when we approximate one image with another. So, a value close to 0 implies a good match and two very different images will have a high Kullback–Leibler divergence, often even higher than 1.
  • Manhattan distance The Minkovski distance is a generalization of the Manhattan distance. Let X and Y be two distributions, the general formula of the Minkovski distance can be expressed as D ( X , Y ) = i X ( i ) Y ( i ) n n . In this paper, the choice is to limit to the first order, also called the 1 norm, corresponding to the Manhattan distance.
We have now explained all the notions necessary for our study. We can now move on to the experiments section.

5. Experimental Comparisons

Before tackling the experimental protocol, here is an example of histogram-based comparisons.

5.1. Example of Histogram-Based Comparisons

Table 3 shows an example of a comparison between several images containing different persons. In most cases, the lowest distance is the distance between the person P 1 at the time t and itself at the time t + x , but sometimes it is not the case. As an example, when the Bhattacharyya distance is used with the CIE L*a*b* color space, the person P 3 at the time t is closer to the person P 1 at the time t + x than the person P 1 itself at the time t. This shows that some associations are better than others and that it is important to choose the color space/distance couple well for better results. That is the purpose of this paper.

5.2. Experimental Protocol

To compare the HOG with different color spaces and the different distance measures, respectively presented in Table 1 and Table 2, a number of videos were used. These videos were all taken with a Basler ace acA2040-120uc (https://www.edmundoptics.com/p/basler-ace-aca2040-120um-monochrome-usb-30-camera/34668/, accessed on 28 March 2022) color camera equipped with a fisheye lens (https://www.edmundoptics.co.uk/p/23quot-format-c-mount-fisheye-lens-18mm-fl/16922/, accessed on 28 March 2022), see Figure 1 for a picture of the device; some features are given in Table 4. The scenario pertains to around 2 to 4 people walking and moving under the camera. The Rotation-Aware People Detection in Overhead fisheye Images (RAPiD) method [12] is used to detect different people in the frames. Indeed, the RAPiD method predicts bounding boxes of people, with a certain center, size, and angle of the bounding box. Even though it is useful for several other tasks, only color features inside the bounding boxes in this study interest us. All the bounding boxes initially in the RGB color space were converted into all the color spaces that are mentioned above. We then computed the different distances between the bounding boxes of the consecutive frames.
For the sake of discussion, we are relying on three videos from our sample; two of them were taken in a hall, as shown in Figure 2 and Figure 4a,b. In the first two videos, there are respectively four and two people walking simultaneously that we want to track. We can see that the RAPiD method is quite robust and detects people, even when the view is quite occluded and people appear unusually smaller in the scene. The third video was taken in a classroom and shows three people walking simultaneously. The camera is much closer this time, and once again, the RAPiD method can detect people, even when they are walking right along the optical axis of the camera, as shown in Figure 4c,d.
However, sometimes, because of a cluttering background or dark lighting, not everyone is detected. As a result, to ensure that all the people were correctly detected, in order to use all the bounding boxes, only the frames where the RAPiD detected the right number of people are kept (k bounding boxes for a video with k segmented persons, with k { 2 , 3 , 4 } ).

5.3. Comparison with Multiple Video Acquisitions

We plotted the graphs corresponding to the similarity scores of the HOG and the different color spaces for a comparison with the four persons in the first video. We have numbered the four persons from zero to three. For a given video frame and for a person i { 0 , 1 , 2 , 3 } , we compute the frame by frame distance between its histogram and the four persons of the following frame. By repeating the experiment on all the frames in the video, we obtain a graph composed of four curves that characterizes a person and the color space/distance or HOG/distance couple. This way, we obtain four graphs per couple for the first video. We repeat the process for the two other videos that have two and three people. A part of these graphs is presented in Figure 5, Figure 6, Figure 7, Figure 8, Figure 9 and Figure 10. We can see that for a given graph, which represents the comparison of a person i ( i { 0 , 1 , 2 , 3 } ), one of the curves is lower than the others, it is the one that represents the comparison of this person i with itself in the following frame. However, we can sometimes observe some spikes, at frames 160 for the graphs comparing persons 1 and 0 (a 1 or b 5 as an example). These spikes can be explained by the intersection of the bounding boxes in some frames. We can observe this in Figure 11, where these bounding boxes are displayed; person 0 appears in the bounding box of person 2 (crossing). Different experiments have been carried out after having compared the people in three different videos, but all the curves tied to the different spaces and with the different similarity measures cannot be displayed due to limited space. The following subsection presents an evaluation of the different comparisons. The goal now is to find the best associations of color space(s), see Table 1 or for the HOG and distance(s) d, see Table 2. Consequently, the data of the curves must be therefore compared between them.

5.4. Evaluating the Comparisons

In this part, the aim is to evaluate which association of the similarity measure with the HOG or color space is the most efficient one to track people. To do so, a metric S is proposed, which is a normalized measure to determine the best among the possible associations.
As mentioned earlier, the underlying objective of this paper is to track people in a video based on the color features. Let us take the example of a video showing two people. Let P 1 t be the person P 1 we want to follow in the frame at time t, and P 2 t be the second person P 2 detected in the same frame. In the next frame, at t + 1 , let P 1 t + 1 and P 2 t + 1 correspond to the detection of P 1 and P 2 , respectively. If the distance between the histograms of P 1 t and P 1 t + 1 is the smallest, then P 1 t + 1 is indeed the person P 1 we want to follow. On the other hand, P 2 t + 1 corresponds to the person P 1 if the distance between the histograms of P 1 t and P 2 t + 1 is the smallest. As a result, the most efficient association of color space and distance is the one that minimizes the distance P 1 P 1 between the histograms of the detection of person P 1 between two frames but maximizes the distance P 1 P 2 between the histograms of the detection of persons P 1 and P 2 between the two frames.
To determine the best association(s), we computed a quantitative score for each video defined as follows for two persons:
S = i n t e r - d i s t a n c e i n t r a - d i s t a n c e = d ^ ( H P 1 H P 1 ) + d ^ ( H P 2 H P 2 ) 2 · d ^ ( H P 1 H P 2 ) ,
where H P i and H P j represent the color histograms of P i and P j , respectively, and d ^ is the average of a distance d (the distances are listed in Table 2) along the video between two consecutive frames. This function can be generalized for n sub-images (targets) present in the video as follows:
S = i n t e r - d i s t a n c e i n t r a - d i s t a n c e = 2 n · ( n 1 ) · ( i , j ) , i j d ^ ( H P i H P i ) + d ^ ( H P j H P j ) 2 · d ^ ( H P i H P j ) .
In this paper, we focused on videos depicting two, three, and four persons so, ( i , j ) { 0 , 1 , 2 , 3 } 2 .
Note that the lower the curve comparing the same two people (from different frames) and the farther away from the others. Consequently, lower will be the score given by the S function implying a better association between the color space and the distance.
Table 5 shows the S scores for all the combinations of color spaces with distances in the video with 4 people. Thus, the similarities between the scores are more visible by distance and not by the color space used. The S score values range between 0.11 and 0.45. Theoretically, the values of the function S could range from 0 to infinity, but in practice, they will very rarely exceed 1. One distance stands out for this video: the KL divergence, as it has a really good score in association with CIE L*a*b and CIE L*u*v* color spaces, but the best association is with YCbCr.
Table 6 shows the S score for all the combinations of color spaces with distances in the video of about two people. Once again, the scores are quite similar when the same distance is computed rather than the same color spaces. S ranges between 0.04 and 0.41. Overall, associations with the correlation distance have a low score; most of them at 0.12, but the Kullback–Leibler (KL) divergence has still the lowest scores. The best associations, in this case, are the YCbCr color space with the KL divergence.
Table 7 shows the S score for all the combinations of color spaces with distances in the video with 3 people. Like for the two other videos, the scores are quite similar when the same distance is computed rather than the same color spaces. S ranges between 0.16 and 0.34. Overall, associations with the correlation, chi-square, and the Kullback–Leibleir divergence have a low score, Ultimately, the association between CIE L*u*v and correlation is the best with a score of 0.09.
One can notice that, compared to the results obtained by using the color spaces, the S scores resulting from the use of the HOG are not that enviable, as a whole. Nevertheless, some scores are even among the best; we must keep in mind that this function S has its limits as it is defined using averages. Consequently, when the overall results of the distance functions are close to 0—this is particularly the case concerning the HOG—the S score is also close to 0. For regular results, it is a good estimator, but as soon as we get more fluctuating results, the S function loses some sense, and this is unfortunately what we can observe in Figure 8, Figure 9 and Figure 10 in the correlation curves, for example.

6. Conclusions

Throughout this paper, the effectiveness of different color spaces and distance pairings for images from a fisheye camera has been evaluated. This study highlights one similarity function, particularly the Kullback–Leibler (KL) divergence. Indeed, this distance obtains the best S score. This can be explained by the fact that its value can be greater than 1; consequently, two very different images can be more easily differentiated. KL divergence works the best with YCbCr space, this combination gives good results in almost all cases, even though it also gives interesting results with CIE L*a*b* and CIE L*u*v* color spaces. Note that some other associations give favorable results such as correlation and chi-square with YCbCr, CIE L*a*b* and CIE L*u*v* or KL divergence with RGB. The HOG can nevertheless obtain very good results, its use is to be seriously considered according to what one seeks; it is necessary nevertheless to keep in mind the limits of the S scores. This study can be useful for works related to tracking. It gives the preferred combination when using a fisheye camera, which is a sensor providing strong deformations in the image. This study was limited to six color spaces and six distances, but it would be interesting to push it further by evaluating more distances and more elaborate color spaces [17]. In addition, we noticed that the RAPiD method was not always 100% accurate since the bounding boxes were sometimes crossed and the background of the detected images could disturb the distance evaluation. It could be interesting to look for and use a more accurate algorithm that would limit the region of interest encompassed by the detected person (its shape) and therefore eliminate the background, optimally. Converting to certain spaces is sometimes time consuming, and so is evaluating distances between two histograms. It would be interesting to see if the reduction(s) of the size of the images or of the histograms cause(s) a loss in precision by bringing a gain in time.

Author Contributions

H.T. and B.M., methodology, resources, software, validation, writing—original draft preparation; M.V., methodology, writing—original draft preparation, writing—review and editing; K.H., formal analysis, investigation, writing—original, draft preparation, writing—review and editing. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study can be found here: https://partage.imt.fr/index.php/s/nytmFqiq8jaztkX (accessed on 28 March 2022).

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

To realize our experiments, we started with images recorded in the RGB color space. In order to test the various spaces of colors, it is necessary to convert these images into the other spaces by the mean of different combinations. Table A1 summarizes the equations which are provided in the OpenCV documentation; they allow passing from RGB space to all the other spaces studied in this paper. Of course, there exist equations computing the inverse and recovering RGB space from the other spaces of colors or passing from any space of color to another, see [17]. We can note in the equations of Table A1 that values of each channel R, G, B are considered to be between 0 and 1. Nevertheless, the inputs are 8-bit images, so the range of each channel is between 0 and 255; it is necessary to simply normalize the different channels as follows:
R = R 255 G = G 255 B = B 255 .
Otherwise, without this normalization, there may be information loss for conversion from linear to non-linear spaces such as for RGB to L*u*v* or L*a*b transformations.
Table A1. Usual color spaces computed from RGB space.
Table A1. Usual color spaces computed from RGB space.
Color SpacesEquation
HSVV m a x ( R , G , B )
S ← V m i n ( R , G , B ) V if V 0 0 otherwise
H ← 60 ( G B ) V m i n ( R , G , B ) if V = R 120 + 60 ( B R ) V m i n ( R , G , B ) if V = G 240 + 60 ( R G ) V m i n ( R , G , B ) if V = B 0 if V = R = G = B
HLS V m a x m a x ( R , G , B )
V m i n m i n ( R , G , B )
L V m a x + V m i n 2
S ← V m a x V m i n V m a x + V m i n if L < 0.5 V m a x V m i n 2 V m a x + V m i n if L 0.5
H ← 60 ( G B ) V m a x V m i n if V m a x = R 120 + 60 ( B R ) V m a x V m i n if V m a x = G 240 + 60 ( R G ) V m a x V m i n if V m a x = B 0 if V m a x = R = G = B
YCbCrY 0.299 R + 0.587 G + 0.114 B
Cr ( R Y ) 0.713 + δ
Cb ( B Y ) 0.564 + δ
where δ = 128 ( In the case of 8 bits images )
CIE L*a*b X Y Z 0.412453 0.357580 0.180423 0.212671 0.715160 0.072169 0.019334 0.119193 0.950227 . R G B
X X X n where X n = 0.950456
Z Z Z n where Z n = 1.088754
L ← 116 Y 1 / 3 16 if Y > 0.008856 903.3 Y if Y 0.008856
a 500 ( f ( X ) f ( Y ) ) + δ
b 200 ( f ( Y ) f ( Z ) ) + δ
where f = t 1 / 3 if t > 0.008856 7.787 t + 16 116 if t 0.008856
and δ = 128 ( In the case of 8 bits images )
CIE L*u*v X Y Z 0.412453 0.357580 0.180423 0.212671 0.715160 0.072169 0.019334 0.119193 0.950227 . R G B
L ← 116 Y 1 / 3 16 if Y > 0.008856 903.3 Y if Y 0.008856
u 4 X X + 15 Y + 3 Z
v 9 Y X + 15 Y + 3 Z
u 13 L ( u u n ) where u n = 0.19793943
v 13 L ( v v n ) where v n = 0.46831096

References

  1. Haggui, O.; Agninoube Tchalim, M.; Magnier, B. A Comparison of OpenCV Algorithms for Human Tracking with a Moving Perspective Camera. In Proceedings of the 2021 9th European Workshop on Visual Information Processing (EUVIP), Paris, France, 3–25 June 2021. [Google Scholar]
  2. Dollar, P.; Wojek, C.; Schiele, B.; Perona, P. Pedestrian detection: An evaluation of the state of the art. IEEE Trans. Pattern Anal. Mach. Intell. 2011, 34, 743–761. [Google Scholar] [CrossRef] [PubMed]
  3. Bosman, H.H.; Petkov, N.; Jonkman, M.F. Comparison of color representations for content-based image retrieval in dermatology. Skin Res. Technol. 2010, 16, 109–113. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  4. Schwarz, M.W.; Cowan, W.B.; Beatty, J.C. An experimental comparison of RGB, YIQ, LAB, HSV, and opponent color models. ACM Trans. Graph. 1987, 6, 123–158. [Google Scholar] [CrossRef]
  5. Velastegui, R.; Pedersen, M. The Impact of Using Different Color Spaces in Histological Image Classification using Convolutional Neural Networks. In Proceedings of the 2021 9th European Workshop on Visual Information Processing (EUVIP), Paris, France, 23–25 June 2021. [Google Scholar]
  6. Zweng, A.; Rittler, T.; Kampel, M. Evaluation of histogram-based similarity functions for different color spaces. In Proceedings of the 14th International Conference on Computer Analysis of Images and Patterns, Seville, Spain, 29–31 August 2011; pp. 455–462. [Google Scholar]
  7. Hansen, P.; Corke, P.; Boles, W. Wide-angle visual feature matching for outdoor localization. Int. J. Robot. Res. 2010, 29, 267–297. [Google Scholar] [CrossRef]
  8. Scaramuzza, D.; Ikeuchi, K. Omnidirectional Camera; Springer: New York, NY, USA, 2014. [Google Scholar]
  9. Dalal, N.; Triggs, B. Histograms of oriented gradients for human detection. In Proceedings of the 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR’05), San Diego, CA, USA, 20–25 June 2005; Volume 1, pp. 886–893. [Google Scholar]
  10. greenHaggui, O.; Bayd, H.; Magnier, B.; Aberkane, A. Human Detection in Moving Fisheye Camera using an Improved YOLOv3 Framework. In Proceedings of the 2021 IEEE 23rd International Workshop on Multimedia Signal Processing (MMSP), Tampere, Finland, 6–8 October 2021. [Google Scholar]
  11. Chiang, S.-H.; Wang, T.; Chen, Y.-F. Efficient pedestrian detection in top-view fisheye images using compositions of perspective view patches. Image Vis. Comput. 2021, 105, 104069. [Google Scholar] [CrossRef]
  12. Duan, Z.; Tezcan, O.; Nakamura, H.; Ishwar, P.; Konrad, J. RAPiD: Rotation-aware people detection in overhead fisheye images. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Seattle, WA, USA, 14–19 June 2020; pp. 636–637. [Google Scholar]
  13. Krams, O.; Kiryati, N. People detection in top-view fisheye imaging. In Proceedings of the 2017 14th IEEE International Conference on Advanced Video and Signal Based Surveillance (AVSS), Lecce, Italy, 29 August–1 September 2017; pp. 1–6. [Google Scholar]
  14. Srisamosorn, V.; Kuwahara, N.; Yamashita, A.; Ogata, T.; Shirafuji, S.; Ota, J. Human position and head direction tracking in fisheye camera using randomized ferns and fisheye histograms of oriented gradients. Vis. Comput. 2019, 36, 1443–1456. [Google Scholar] [CrossRef]
  15. greenHaggui, O.; Vert, M.; McNamara, K.; Brieussel, B.; Magnier, B. Human Tracking in Top-view Fisheye Images with Color Histograms via Deep Learning Detection. In Proceedings of the 2021 IEEE International Conference on Imaging Systems and Techniques (IST), Kaohsiung, Taiwan, 24–26 August 2021. [Google Scholar]
  16. Pratt, W.K. Digital Image Processing, 4th ed.; John Wiley & Sons, Inc.: Hoboken, NJ, USA, 2007. [Google Scholar]
  17. Ibraheem, N.A.; Hasan, M.M.; Khan, R.Z.; Mishra, P.K. Understanding color models: A review. ARPN J. Sci. Technol. 2012, 2, 265–275. [Google Scholar]
  18. Magnier, B.; Abdulrahman, H.; Montesinos, P. A review of supervised edge detection evaluation methods and an objective comparison of filtering gradient computations using hysteresis. J. Imaging 2018, 4, 74. [Google Scholar] [CrossRef] [Green Version]
  19. Zhang, L.; Van Der Maaten, L. Preserving structure in model-free tracking. IEEE Trans. Pattern Anal. Mach. Intell. 2013, 36, 756–769. [Google Scholar] [CrossRef] [PubMed]
  20. Comaniciu, D.; Ramesh, V.; Meer, P. Kernel-based object tracking. IEEE Trans. Pattern Anal. Mach. Intell. 2003, 25, 564–577. [Google Scholar] [CrossRef] [Green Version]
Figure 1. Basler ace acA2040-120uc color camera equipped with a fish eye lens utilized in our experiments. It is positioned next to a EUR 0.05 coin.
Figure 1. Basler ace acA2040-120uc color camera equipped with a fish eye lens utilized in our experiments. It is positioned next to a EUR 0.05 coin.
Jimaging 08 00115 g001
Figure 2. Persons extracted from fisheye images and displaying of different color space representation with their tied histograms.
Figure 2. Persons extracted from fisheye images and displaying of different color space representation with their tied histograms.
Jimaging 08 00115 g002
Figure 3. HOG computation on a full gray-scale image using [ 1 0 1 ] and [ 1 0 1 ] t masks.
Figure 3. HOG computation on a full gray-scale image using [ 1 0 1 ] and [ 1 0 1 ] t masks.
Jimaging 08 00115 g003
Figure 4. Extracted frames from videos 2 and 3 with two and three persons respectively. (a) Video 2, frame 194. (b) Video 2, frame 494. (c) Video 3, frame 94. (d) Video 3, frame 192.
Figure 4. Extracted frames from videos 2 and 3 with two and three persons respectively. (a) Video 2, frame 194. (b) Video 2, frame 494. (c) Video 3, frame 94. (d) Video 3, frame 192.
Jimaging 08 00115 g004
Figure 5. Curves comparing persons from one frame to another in video 1 (persons 0, 1, 2, and 3). (a) RGB and Correlation. (b) HLS and Intersection. (c) HSV and Chi-square. (d) L*a*b and Bhattacharyya. (e) L*u*v and KL divergence. (f) YCrCb and Minkovski.
Figure 5. Curves comparing persons from one frame to another in video 1 (persons 0, 1, 2, and 3). (a) RGB and Correlation. (b) HLS and Intersection. (c) HSV and Chi-square. (d) L*a*b and Bhattacharyya. (e) L*u*v and KL divergence. (f) YCrCb and Minkovski.
Jimaging 08 00115 g005
Figure 6. Curves comparing persons from one frame to another in video 2 (persons 0, 1 and 2).
Figure 6. Curves comparing persons from one frame to another in video 2 (persons 0, 1 and 2).
Jimaging 08 00115 g006
Figure 7. Curves comparing persons from one frame to another in video 3 (persons 0 and 1).
Figure 7. Curves comparing persons from one frame to another in video 3 (persons 0 and 1).
Jimaging 08 00115 g007
Figure 8. Curves comparing persons from one frame to another in video 1 (persons 0, 1, 2, and 3) using the HOG. (a) Correlation. (b) Intersection. (c) Chi-square. (d) Bhattacharyya. (e) KL divergence. (f) Minkovski.
Figure 8. Curves comparing persons from one frame to another in video 1 (persons 0, 1, 2, and 3) using the HOG. (a) Correlation. (b) Intersection. (c) Chi-square. (d) Bhattacharyya. (e) KL divergence. (f) Minkovski.
Jimaging 08 00115 g008
Figure 9. Curves comparing persons from one frame to another in video 2 (persons 0, 1, and 2) using the HOG.
Figure 9. Curves comparing persons from one frame to another in video 2 (persons 0, 1, and 2) using the HOG.
Jimaging 08 00115 g009
Figure 10. Curves comparing persons from one frame to another in video 3 (persons 0 and 1) using the HOG.
Figure 10. Curves comparing persons from one frame to another in video 3 (persons 0 and 1) using the HOG.
Jimaging 08 00115 g010
Figure 11. Extractions of the four persons from the first video (namely P i , i { 0 , 1 , 2 , 3 } ) and extractions where there is more than one person (bottom). Note that images are unwrapped for displaying and of different sizes.
Figure 11. Extractions of the four persons from the first video (namely P i , i { 0 , 1 , 2 , 3 } ) and extractions where there is more than one person (bottom). Note that images are unwrapped for displaying and of different sizes.
Jimaging 08 00115 g011
Table 1. Typical color spaces with their acronyms and range in OpenCV software for histogram computation.
Table 1. Typical color spaces with their acronyms and range in OpenCV software for histogram computation.
Space’s NameComponentsRange
R (Red)[0, 255]
RGBG (Green)[0, 255]
B (Blue)[0, 255]
H (Hue)[0, 180]
HSVS (Saturation)[0, 255]
V (Value)[0, 255]
H (Hue)[0, 180]
HLSL (Lightness)[0, 255]
S (Saturation)[0, 255]
L* (Lightness)[0, 255]
CIE L*a*b*a* (Red/green coordinate)[0, 255]
b* (Yellow/blue coordinate)[0, 255]
L* (Lightness)[0, 255]
CIE L*u*v*u* (Red/green coordinate)[0, 255]
v* (Yellow/blue coordinate)[0, 255]
Y (Luminance)[0, 255]
YCbCrCr (Red-difference)[0, 255]
Cb (Blue-difference)[0, 255]
Table 2. Distance measures to compare 2 different histograms H 1 and H 2 of the same length. Usually, H 1 and H 2 are both normalized.
Table 2. Distance measures to compare 2 different histograms H 1 and H 2 of the same length. Usually, H 1 and H 2 are both normalized.
Distance MetricsEquation
Correlation d ( H 1 , H 2 ) = 1 2 · 1 k = 0 N 1 ( H 1 ( k ) H 1 ¯ ) · ( H 2 ( k ) H 2 ¯ ) k = 0 N 1 ( H 1 ( k ) H 1 ¯ ) 2 · k = 0 N 1 ( H 2 ( k ) H 2 ¯ ) 2
Chi-square χ 2 d ( H 1 , H 2 ) = k = 0 N 1 ( H 1 ( k ) H 2 ( k ) ) 2 2 · ( H 1 ( k ) + H 2 ( k ) )
Intersection d ( H 1 , H 2 ) = 1 k = 0 N 1 m i n ( H 1 ( k ) , H 2 ( k ) )
Bhattacharyya d ( H 1 , H 2 ) = 1 1 H 1 ¯ · H 2 ¯ · N 2 k = 0 N 1 H 1 ( k ) · H 2 ( k )
KL divergence d ( H 1 , H 2 ) = k = 0 N 1 H 1 ( k ) · l o g H 1 ( k ) H 2 ( k )
Manhattan distance d ( H 1 , H 2 ) = k = 0 N 1 | H 1 ( k ) H 2 ( k ) |
Table 3. Example of comparison of distance scores between several persons in two different frames as a function of the color space. The reference image on the third line is extracted in the image presented in Figure 2b whereas other persons are tied to the detected persons in Figure 2a. Note that images are unwrapped for displaying.
Table 3. Example of comparison of distance scores between several persons in two different frames as a function of the color space. The reference image on the third line is extracted in the image presented in Figure 2b whereas other persons are tied to the detected persons in Figure 2a. Note that images are unwrapped for displaying.
Jimaging 08 00115 i001 Person P1 at Time t + x
Image at Time tDistanceRGBHSVHLSL*a*b*L*u*v*YCrCbHOG
Person P0Correlation0.270.140.130.090.090.10.48
    Jimaging 08 00115 i002Chi-square χ 2 0.20.170.150.180.180.20.03
Intersection0.380.320.310.330.330.350.1
Bhattacharyya0.340.320.30.350.350.370.08
Kullback–Leibler0.580.510.430.440.470.50.05
Minkovski0.760.650. 630. 660.720.70.19
Person P1Correlation0.180.080.070.050.050.050.45
    Jimaging 08 00115 i003Chi-square χ 2 0.140.090.090.10.10.10.06
Intersection0.280.220.220.220.220.220.18
Bhattacharyya0.310.250.240.260.260.270.18
Kullback–Leibler0.380.290.30.250.290.260.15
Minkovski0.570.450.450.430.470.450.36
Person P2Correlation0.270.10.110.050.050.060.7
    Jimaging 08 00115 i004Chi-square χ 2 0.190.120.130.110.110.120.04
Intersection0.350.260.270.240.240.260.14
Bhattacharyya0.340.290.290.260.270.270.15
Kullback–Leibler0.650.660.80.330.430.380.09
Minkovski0.710. 530.540.470.540.530.27
Person P3Correlation0.240.080.080.040.040.050.6
    Jimaging 08 00115 i005Chi-square χ 2 0.180.110.10.090.090.10.08
Intersection0.360.260.240.230.230.250.22
Bhattacharyya0.330.260.260.230.230.240.25
Kullback–Leibler0.670.480.640.290.410.331.53
Minkovski0.730.530.490.460.50.490.45
Table 4. Main characteristics of the applied database.
Table 4. Main characteristics of the applied database.
Number of ImagesFrame RateWhereNumber of People to Track
Video 153540Hall4
Video 253540Hall2
Video 353540Classroom3
Table 5. S scores calculated for the 1st video with 4 people.
Table 5. S scores calculated for the 1st video with 4 people.
Color Spaces or DescriptorCorrelationChi-Square ( χ 2 ) IntersectionBhattacharyyaKullback–LeiblerManhattan
RGB0.140.160.360.390.160.38
HSV0.180.220.420.450.220.44
HLS0.170.220.410.450.220.43
CIE L*a*b*0.160.140.350.350.120.36
CIE L*u*v*0.150.140.340.350.120.35
YCbCr0.140.130.330.340.110.35
HOG0.170.160.340.350.160.34
Table 6. S scores calculated for the 2nd video with 2 people.
Table 6. S scores calculated for the 2nd video with 2 people.
Color Spaces or DescriptorCorrelationChi-Square ( χ 2 ) IntersectionBhattacharyyaKullback–LeiblerManhattan
RGB0.120.140.340.350.110.34
HSV0.120.180.380.400.140.38
HLS0.120.190.380.410.150.38
CIE L*a*b*0.140.130.330.320.060.33
CIE L*u*v*0.120.120.310.310.050.31
YCbCr0.120.110.300.300.040.30
HOG0.110.210.590.600.200.59
Table 7. S scores calculated for the 3rd video with 3 people.
Table 7. S scores calculated for the 3rd video with 3 people.
Color Spaces or DescriptorCorrelationChi-Square ( χ 2 ) IntersectionBhattacharyyaKullback–LeiblerManhattan
RGB0.200.210.310.330.160.32
HSV0.200.210.330.340.190.33
HLS0.200.210.330.350.200.33
CIE L*a*b*0.190.200.300.320.180.30
CIE L*u*v*0.190.190.290.300.170.29
YCbCr0.190.200.290.310.170.29
HOG0.130.120.270.280.120.27
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Talaoubrid, H.; Vert, M.; Hayat, K.; Magnier, B. Human Tracking in Top-View Fisheye Images: Analysis of Familiar Similarity Measures via HOG and against Various Color Spaces. J. Imaging 2022, 8, 115. https://doi.org/10.3390/jimaging8040115

AMA Style

Talaoubrid H, Vert M, Hayat K, Magnier B. Human Tracking in Top-View Fisheye Images: Analysis of Familiar Similarity Measures via HOG and against Various Color Spaces. Journal of Imaging. 2022; 8(4):115. https://doi.org/10.3390/jimaging8040115

Chicago/Turabian Style

Talaoubrid, Hicham, Marina Vert, Khizar Hayat, and Baptiste Magnier. 2022. "Human Tracking in Top-View Fisheye Images: Analysis of Familiar Similarity Measures via HOG and against Various Color Spaces" Journal of Imaging 8, no. 4: 115. https://doi.org/10.3390/jimaging8040115

APA Style

Talaoubrid, H., Vert, M., Hayat, K., & Magnier, B. (2022). Human Tracking in Top-View Fisheye Images: Analysis of Familiar Similarity Measures via HOG and against Various Color Spaces. Journal of Imaging, 8(4), 115. https://doi.org/10.3390/jimaging8040115

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop