You are currently viewing a new version of our website. To view the old version click .
Applied Sciences
  • Article
  • Open Access

16 December 2021

Time Classification Algorithm Based on Windowed-Color Histogram Matching

,
and
Department of IT Engineering, Sookmyung Women’s University, 100 Chungpa-ro 47 gil, Yongsna-gu, Seoul 04310, Korea
*
Author to whom correspondence should be addressed.
This article belongs to the Topic Applied Computer Vision and Pattern Recognition

Abstract

A web-based search system recommends and gives results such as customized image or video contents using information such as user interests, search time, and place. Time information extracted from images can be used as a important metadata in the web search system. We present an efficient algorithm to classify time period into day, dawn, and night when the input is a single image with a sky region. We employ the Mask R-CNN to extract a sky region. Based on the extracted sky region, reference color histograms are generated, which can be considered as the ground-truth. To compare the histograms effectively, we design the windowed-color histograms (for RGB bands) to compare each time period from the sky region of the reference data with one of the input images. Also, we use a weighting approach to reflect a more separable feature on the windowed-color histogram. With the proposed windowed-color histogram, we verify about 91% of the recognition accuracy in the test data. Compared with the existing deep neural network models, we verify that the proposed algorithm achieves better performance in the test dataset.

1. Introduction

Generally, we design a web-based search system to get accurate answers for the given queries. Meanwhile, the type of search query has diversified in web and mobile environments. In the past, only keywords were used as input query but nowadays it has been extended to sentence, voice, image, and video [1,2,3,4]. In addition, by expanding the definition of semantic search, the searching service is developing into a customized service that not only shows objective results but also recommends information that users may like [5]. For example, they recommend and show customized image or video content using information such as user interests, search time, climate, and the place where the user resides.
To use this function, it is essential to use some additional contextual information. Researchers has actively conducted to apply deep learning to extract context information from text, image, video, and voice data. In particular, by using contextual information based on image or video data, it has so far been successfully applied to computer vision tasks such as object detection, semantic segmentation, and image classification [6,7,8,9,10,11,12]. This information can describe objects’ class, background, as well as some situations or relationship between them [13,14,15].
There are some useful components in image and video processing. For a long time, texture and color components have had low-level key features for scene description [15]. By using this information, an image descriptor called Contextual Mean Census Transform (CMCT), which combines distribution of local structures and contextual information, was designed to classify scenes. However, it needs to understand a more high-level context, such as the relationship between people or their emotions in the scene. This means that the traditional feature-based approaches cannot support making high-level understanding. Therefore, more high-level abstractive information such as when (time), where (location), what to do (action), with whom (relationship with other) and so on should be extracted from the image and video data [16,17,18].
In order to obtain high-level abstracted information, we focus on extracting time information from images. It can be used as important metadata in the search or recommendation system. To extract it from image data, we paid attention to the amount of sunlight because it causes the change of time during the day—dawn, morning, midday, evening, and night. In this aspect, the sky region represents the change of time in outdoor images well and can provide useful information about the environment, especially for the given time. Therefore, by extracting sky area and analyzing its characteristics, it is able to derive meaningful time information during the day.
In this paper, we propose an efficient algorithm to classify time by analyzing the sky region with deep learning and color histogram. The proposed algorithm is trained by Mask R-CNN [19] to extract the sky region from the given image. Based on the extracted region, the reference color histograms are generated, which can be considered as the ground-truth. Time is classified by comparing the reference color histogram with a histogram of the sky region extracted from test dataset. We design the windowed-color histograms (for RGB bands) to compare each time period with the reference and input efficiently. In addition, we propose a weighted histogram representation which compares histograms by assigning large weights according to the degree of color importance.
The main contribution of this study is the followings:
  • We focus on the sky area of the image because it has more information on light and discriminating feature;
  • The designed weighted histogram comparison can pay attention on more important colors in each class;
  • As a result, we simplify the problem and improve the time classification performance better than existing deep learning models.
This study can have a wide range of applications in the real world. If time information is used, a web-based search system can categorize results into dawn, dusk, day, or night. In addition, it can recommend contents based on the related time of queries. We aim to provide more accurate search and recommendation results by using time information of image or video.
The remaining part of this paper is organized as follows: In Section 2, we analyze some existing methods as the related works. Section 3 describes the proposed approach and Section 4 describes the experimental results. Conclusions are presented in Section 5.

3. Proposed Time Classification Algorithm

Figure 1 shows the overall procedure of the proposed time classification algorithm based on the sky information. In the training step, the dataset including the sky is trained using Mask R-CNN to detect the sky region. A reference color histogram representing the time period (dawn, day, and night) is created based on the sky information. In the evaluation step, the sky region is extracted from test image through the trained Mask R-CNN and a color histogram of the sky area is generated. Finally, the test image is classified into dawn, day, and night in the comparison process between the color histogram of the test image and the reference color histogram.
Figure 1. The flow chart of the proposed algorithm.

3.1. Sky Detection

For sky region detection, the Mask R-CNN [19] is trained by using a dataset consisted of images and annotations of the sky region. The Mask R-CNN is well known model in semantic segmentation task and performs well even in small dataset during transfer learning. The model is initialized using the pre-trained MS COCO weight. Additionally, we freeze the head of Mask R-CNN to extract the generalized feature because the weight of the other layers can be trained to fit the sky detection dataset. Figure 2 shows the results of the detected sky region.
Figure 2. The results of the detected sky region using the Mask R-CNN [19].

3.2. Time Classification

A reference color histogram representing each time period is generated based on only the sky region extracted from the training dataset. We use four methods to compare the histogram generated from test image with the reference color histogram: Correlation, Intersection, Bhattacharyya, and Chi-square. We find the optimal comparison method and parameters among them. We apply a weighted sum comparison between the color histograms.

3.2.1. Windowed-Color Histogram

With the average of the annotated sky region in train dataset, we construct the reference color histogram of each time period. Figure 3 shows the histograms of dawn, day, and night class. In windowed-color histogram concept, we consider some dominant color parts more when compared with each other. As shown in Figure 4, we take a limited range (limited bandwidth) of the histogram centered on the average value of each color band.
Figure 3. Reference color histograms (B, G, R bands) of each time period. (a) Reference color histogram of “dawn” or “dusk” time periods. (b) Reference color histogram of “day” time period. (c) Reference color histogram of “night” time period.
Figure 4. The windowed-based color histograms of the generated reference color histograms. (a) “dawn” or “dust” case. (b) “day” case. (c) “night” case.
The windowed-color histogram of input image is computed with the pre-defined bandwidth. We compute the matching score with the windowed-color histograms of the references to classify the input image into each time period class, as shown in Equation (1).
d i ( H 1 i , H 2 i ) = F ( H 1 { B W 2 , B W 2 } i , H 2 { B W 2 , B W 2 } i ) ,
where F ( · ) is one of the histogram comparing methods, B W is a pre-defined range centered on the average value of each color, and i is an index of {blue, green, and red} bands.

3.2.2. Histogram Comparison Methods

  • Correlation
Correlation is a statistical technique that shows the relationship between two variables, whether it is linear or non-linear and how strongly they are related. The result of a correlation is called the correlation coefficient and the equation of it is:
d ( H 1 , H 2 ) = I ( H 1 ( I ) H ¯ 1 ) ( H 2 ( I ) H ¯ 2 ) I ( H 1 ( I ) H ¯ 1 ) 2 I ( H 2 ( I ) H ¯ 2 ) 2 ,
where I is an index of histogram bins and H 1 ( I ) and H 2 ( I ) are the histograms. In addition H ¯ k is the average value of intensity. The denominator of this equation stands for the multiple of the standard deviation of H 1 ( I ) and H 2 ( I ) . The numerator stands for the covariance of H 1 ( I ) and H 2 ( I ) , so it means the change between H 1 ( I ) and H 2 ( I ) . The possible value of the correlation lies between −1 and 1. If two histograms are closely related, d ( H 1 , H 2 ) is closer to +1 or −1. If not related, then d ( H 1 , H 2 ) is closer to 0.
  • Intersection
Intersection calculates the similarity of two probability distributions or histograms [38]. The equation is:
d ( H 1 , H 2 ) = I m i n ( H 1 ( I ) , H 2 ( I ) ) ,
where I is an index of histogram bins and H 1 ( I ) and H 2 ( I ) are the histograms. The possible value of the intersection lies between 0 and 1. If two histograms are closely related, d ( H 1 , H 2 ) becomes a larger number. If not related, then d ( H 1 , H 2 ) is closer to 0, which means there is no overlap.
  • Bhattacharyya
Bhattacharyya distance measures the similarity of two probability distributions or histograms [39]. The Bhattacharyya coefficient is a measure of the amount how two probability distributions or histograms are overlapped. The equation of Bhattacharyya coefficient is:
B C ( H 1 , H 2 ) = I H 1 ( I ) · H 2 ( I ) ,
where I is an index of histogram bins and H 1 ( I ) and H 2 ( I ) are the histograms. As you can see, the Bhattacharyya coefficient stands for the sum of discrete probability distribution, which is overlapped between two probability distributions or histograms and the range is [0,1]. If two histograms are closely related, d ( H 1 , H 2 ) is closer to 1. d ( H 1 , H 2 ) will be 0, if there is no overlap at all. In fact, Hellinger distance can be defined as:
d ( H 1 , H 2 ) = 1 B C ( H 1 , H 2 ) .
According to the Bhattacharyya coefficient, if two histograms are closely related, d ( H 1 , H 2 ) is closer to 0. If not related, then d ( H 1 , H 2 ) is closer to 1.
  • Chi-Square
Chi-square distance is one of the distance measures that can be used to find dissimilarity between two histograms [40]. The expression can be defined as:
d ( H 1 , H 2 ) = I ( H 1 ( I ) H 2 ( I ) ) 2 H 1 ( I ) ,
where I is an index of histogram bins and H 1 ( I ) and H 2 ( I ) are the histograms. It means the difference against the denominator histograms. If two histograms are closely related, d ( H 1 , H 2 ) becomes 0, because H 1 ( I ) H 2 ( I ) becomes 0. If not related, then d ( H 1 , H 2 ) becomes a larger number.
To summarize, Intersection and Bhattacharyya methods are techniques used to measure the similarity of two histograms by calculating the overlap area. Therefore, these methods are effective for comparing histograms that have different color features.

3.2.3. Weighted Histogram Comparison

Brightness of the acquired image depends on the color band during the day. We observe that some specific color components are important to separate the time period. For example, blue and green components can give good separability between “dawn” and “day” periods. However, a red color is not a good indicator to verify them, as shown in Figure 3. Therefore, we design a weighted matching scheme between the reference color histogram and the input image histogram as the following:
d ( H 1 , H 2 ) = i w i d i ( H 1 i , H 2 i )
where d i is defined as the i-th band difference in the above, w i is a tuned weighting factor for each color band. We set this weighting factor w i through the experiments.
With the above weighted histogram mechanism, we use a hierarchical classification structure: (1) First, the input image is classified into “day (dawn)”, and “night” classes based on the windowed-color histogram and weighting factor, (2) After that, the input image, which is in the "day (dawn)" class, is separated into “day” and “dawn” classes based on the refined windowed-color histogram and weighting factor.

4. Experimental Results

4.1. Experimental Environment

The proposed algorithm has been implemented in Python language. We used i7 CPU with GeForce GTX 1070 GPU (8 G Byte) and 16 GB RAM. Deep learning frameworks for extracting sky area are Tensorflow 1.14.0 and Keras 2.1.0. For training the Mask R-CNN, we used Stochastic gradient descent (SGD) as optimizer and set the same parameters (learning rate = 0.001 and 30 epochs) in [38].

4.2. Dataset

The dataset for time classification used a total of 450 images and divided them to 150 images for the training dataset, 300 images for the validation dataset, and 300 images for the test dataset. Images are collected by web crawling and various time scenes are captured from movies. The train and validation dataset are used to extract the sky region using Mask R-CNN. They consist of images, annotations of sky regions, and are classified by time variance: dusk or dawn, day, night. The validation dataset is used only for sky detection using the Mask R-CNN. After training the Mask R-CNN model, the reference histogram of each time class is created to compare the color and brightness of the sky region. Figure 5 shows a part of our dataset for each time period.
Figure 5. Example of each time period from the dataset.

4.3. Results and Discussion

4.3.1. Sky Detection

With the validation dataset, we compare the detection accuracy between using Shen’s method [32] and the Mask R-CNN method. The detection accuracy is shown in Table 1. The Mask R-CNN achieved 91.67% accuracy. It has much better performance than Shen’s method. From this result, we can see the the Mask R-CNN method is very powerful to separate the desired object from the natural background.
Table 1. The accuracy comparison between Shen’s method [32] and the Mask R-CNN method for sky detection.

4.3.2. Time Classification

We first classify “night” and “day” using a hierarchical classification structure to simplify the problem. Intersection and Battacharya histogram comparison methods are appropriate to classify images with distinct features. Then, the bandwidth ( B W ) is changed according to the average of each channel of the reference color histogram. As a result of the experiment, we find that the B W = 120 and intersection histogram comparison method perform best when classifying images into “night” and “day (dawn)”. As mentioned previously, the difference in contribution of each color component (BGR) should be considered.
Table 2 shows the result of each color contribution experiment calculated by the intersection measure. According to the intersection result, green is more important than other colors. Therefore, we calculate the accuracy using the weighted histograms comparison by assigning the largest weight to green. In Table 3, when the weight ratio of blue, green, and red was { 0 : 6 : 4 } and the time classification accuracy was 98.33%, we achieved the best results.
Table 2. Result of classifying “night” and “day (dawn)” time periods using intersection with bandwidth = 120.
Table 3. Result of classifying “night” and “day (dawn)” time periods using intersection with the weighting approach.
After classifying “night” time, the remaining dataset is classified as “dawn” or “day”. Similar to the “night” time classification, the bandwidth was adjusted using a histogram comparison method. As a result, the correlation method performed better than the other methods, including intersection, and it was effective to consider all sections of B W .
In Table 4, as a result of calculating the importance of each histogram color in the correlation, the blue color is important to distinguish between day and dawn. It means that the blue color has good separable feature to classify the sky region into “dawn” or “dusk” and “day” time periods. On the other hand, the red color affects less than the blue color. The green color band almost does not affect this classification. Therefore, we assign the largest weight to the blue color band and an accuracy of 91% was achieved when the blue, red, green were weighted by w i = { 6 : 1 : 3 } , respectively as shown in Table 5.
Table 4. Result of classifying “dawn” or “dusk” and “day” time periods using the correlation method with all range of the bandwidth.
Table 5. Result of classifying “dawn” or “dusk” and “day” time periods using the correlation method with the weighting approach.
Table 6 shows the performance comparison between the proposed method and the existing deep learning classification models. We employed ResNet [6], DenseNet [25], XceptionNet [26], InceptionResNet [27], NasNet [28], and EfficientNet [29]. The same train and validation dataset were used, and the classification performance was inferred from the test dataset. The deep learning models obtained the result with transfer learning using the pre-trained ImageNet weights. Among them, ResNet152 achieved the highest accuracy of 87.67%. The proposed algorithm achieved 91% of accuracy as shown bold face. This means that the proposed scheme is superior to the existing deep learning models for time period classification even though three classes (day, dawn, and night) exist.
Table 6. Performance comparison of the proposed algorithm with the existing deep learning classification models.
The existing models have high complexity because they train a whole area of images and find optimal feature. The usual dataset includes various place, season, objects, and backgrounds. Therefore, it needs to focus on the discriminative feature of time information. We simplify this problem through three approaches: sky detection, weighted histogram comparison, and hierarchical classification structure. As a result, we have better performance than other models. However, there are a few things to that need to be improved. If various data can be collected, we can generate a more representative reference histogram for robust classification. In addition, future studies will be needed for more detailed time classification such as dawn, morning, noon, afternoon, evening, and deep night.

5. Conclusions

This paper has proposed an efficient time-classification algorithm using the sky region detection approach based on a single image. The sky region of an image contains a change in brightness over time. For sky detection, we trained the Mask R-CNN model and made reference color histograms of all time periods (day, dawn, and night). We employed two measures for the proposed windowed-color histogram. Additionally, we designed the weighting factor of each channel to reflect the characteristics of each color and made more efficient classifications. Based on the proposed algorithm, we have achieved 91% of the verification accuracy with our hierarchical classification structure.

Author Contributions

Conceptualization, B.-G.K.; methodology, B.-G.K. and H.-J.P.; software, H.-J.P.; validation, J.-I.J. and H.-J.P.; formal analysis, B.-G.K. and H.-J.P.; investigation, J.-I.J. and H.-J.P.; writing—original draft preparation, J.-I.J. and H.-J.P.; writing—review and editing, B.-G.K.; supervision, B.-G.K.; project administration, B.-G.K. All authors have read and agreed to the published version of the manuscript.

Funding

This research received funding from the Ministry of Culture, Sports and Tourism (MCST) and from the Korea Copyright Commission in 2021.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

This research project was supported by Ministry of Culture, Sports, and Tourism (MCST) and from the Korea Copyright Commission in 2021.

Conflicts of Interest

The author declares no conflict of interest.

References

  1. Shiyamala, M.; Kalaiarasi, G. Contextual image search with keyword and image input. In Proceedings of the International Conference on Information Communication and Embedded Systems (ICICES2014), Chennai, India, 27–28 February 2014; pp. 1–5. [Google Scholar]
  2. Krapac, J.; Allan, M.; Verbeek, J.; Juried, F. Improving web image search results using query-relative classifiers. In Proceedings of the 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, San Francisco, CA, USA, 13–18 June 2010; pp. 1094–1101. [Google Scholar]
  3. Guy, I. The characteristics of voice search: Comparing spoken with typed-in mobile web search queries. ACM Trans. Inf. Syst. 2018, 36, 1–28. [Google Scholar] [CrossRef]
  4. Guy, I. Searching by talking: Analysis of voice queries on mobile web search. In Proceedings of the 39th International ACM SIGIR conference on Research and Development in Information Retrieval, Pisa, Italy, 17–21 July 2016; pp. 35–44. [Google Scholar]
  5. Eom, W.; Lee, S.; De Neve, W.; Ro, Y.M. Improving image tag recommendation using favorite image context. In Proceedings of the 2011 18th IEEE International Conference on Image Processing, Brussels, Belgium, 11–14 September 2011; pp. 2445–2448. [Google Scholar]
  6. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  7. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  8. Jeong, D.; Kim, B.G.; Dong, S.Y. Deep joint spatiotemporal network (DJSTN) for efficient facial expression recognition. Sensors 2020, 20, 1936. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  9. Kim, J.H.; Kim, B.G.; Roy, P.P.; Jeong, D.M. Efficient facial expression recognition algorithm based on hierarchical deep neural network structure. IEEE Access 2019, 7, 41273–41285. [Google Scholar] [CrossRef]
  10. Kim, J.H.; Hong, G.S.; Kim, B.G.; Dogra, D.P. deepGesture: Deep learning-based gesture recognition scheme using motion sensors. Displays 2018, 55, 38–45. [Google Scholar] [CrossRef]
  11. Yang, Z.; Leng, L.; Kim, B.G. StoolNet for color classification of stool medical images. Electronics 2019, 8, 1464. [Google Scholar] [CrossRef] [Green Version]
  12. Jang, S.; Battulga, L.; Nasridinov, A. Detection of Dangerous Situations using Deep Learning Model with Relational Inference. J. Multimed. Inf. Syst. 2020, 7, 205–214. [Google Scholar] [CrossRef]
  13. Pedronette, D.C.G.; Torres, R.d.S. Exploiting contextual information for image re-ranking and rank aggregation. Int. J. Multimed. Inf. Retr. 2012, 1, 115–128. [Google Scholar] [CrossRef] [Green Version]
  14. Song, X.B.; Abu-Mostafa, Y.; Sill, J.; Kasdan, H.; Pavel, M. Robust image recognition by fusion of contextual information. Inf. Fusion 2002, 3, 277–287. [Google Scholar] [CrossRef] [Green Version]
  15. de Souza Gazolli, K.A.; Salles, E.O.T. A contextual image descriptor for scene classification. Online Proc. Trends Innov. Comput. 2012, 66–71. [Google Scholar]
  16. Pazzani, M.J.; Billsus, D. Content-based recommendation systems. In The Adaptive Web; Springer: Berlin/Heidelberg, Germany, 2007; pp. 325–341. [Google Scholar]
  17. Lops, P.; De Gemmis, M.; Semeraro, G. Content-based recommender systems: State of the art and trends. In Recommender Systems Handbook; Springer: Berlin/Heidelberg, Germany, 2011; pp. 73–105. [Google Scholar]
  18. Popescul, A.; Ungar, L.H.; Pennock, D.M.; Lawrence, S. Probabilistic models for unified collaborative and content-based recommendation in sparse-data environments. arXiv 2013, arXiv:1301.2303. [Google Scholar]
  19. He, K.; Gkioxari, G.; Dollár, P.; Girshick, R. Mask r-cnn. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2961–2969. [Google Scholar]
  20. Saha, B.; Davies, D.; Raghavan, A. Day Night Classification of Images Using Thresholding on HSV Histogram. U.S. Patent 9,530,056, 27 December 2016. [Google Scholar]
  21. Taha, M.; Zayed, H.H.; Nazmy, T.; Khalifa, M. Day/night detector for vehicle tracking in traffic monitoring systems. Int. J. Comput. Electr. Autom. Control. Inf. Eng. 2016, 10, 101. [Google Scholar]
  22. Park, K.H.; Lee, Y.S. Classification of Daytime and Night Based on Intensity and Chromaticity in RGB Color Image. In Proceedings of the 2018 International Conference on Platform Technology and Service (PlatCon), Jeju, Korea, 29–31 January 2018; pp. 1–6. [Google Scholar]
  23. Shen, J.; Robertson, N. BBAS: Towards large scale effective ensemble adversarial attacks against deep neural network learning. Inf. Sci. 2021, 569, 469–478. [Google Scholar] [CrossRef]
  24. Shi, Y.; Wei, Z.; Ling, H.; Wang, Z.; Zhu, P.; Shen, J.; Li, P. Adaptive and robust partition learning for person retrieval with policy gradient. IEEE Trans. Multimed. 2020, 23, 3264–3277. [Google Scholar] [CrossRef]
  25. Huang, G.; Liu, Z.; Van Der Maaten, L.; Weinberger, K.Q. Densely connected convolutional networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 4700–4708. [Google Scholar]
  26. Chollet, F. Xception: Deep learning with depthwise separable convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 1251–1258. [Google Scholar]
  27. Szegedy, C.; Ioffe, S.; Vanhoucke, V.; Alemi, A.A. Inception-v4, inception-resnet and the impact of residual connections on learning. In Proceedings of the Thirty-First AAAI Conference on Artificial Intelligence, San Francisco, CA, USA, 4–9 February 2017. [Google Scholar]
  28. Zoph, B.; Vasudevan, V.; Shlens, J.; Le, Q.V. Learning transferable architectures for scalable image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 8697–8710. [Google Scholar]
  29. Tan, M.; Le, Q. Efficientnet: Rethinking model scaling for convolutional neural networks. In Proceedings of the International Conference on Machine Learning. PMLR, Long Beach, CA, USA, 9–15 June 2019; pp. 6105–6114. [Google Scholar]
  30. Volokitin, A.; Timofte, R.; Van Gool, L. Deep features or not: Temperature and time prediction in outdoor scenes. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Las Vegas, NV, USA, 25 June–1 July 2016; pp. 63–71. [Google Scholar]
  31. Ibrahim, M.R.; Haworth, J.; Cheng, T. WeatherNet: Recognising weather and visual conditions from street-level images using deep residual learning. ISPRS Int. J. Geo-Inf. 2019, 8, 549. [Google Scholar] [CrossRef] [Green Version]
  32. Shen, Y.; Wang, Q. Sky region detection in a single image for autonomous ground robot navigation. Int. J. Adv. Robot. Syst. 2013, 10, 362. [Google Scholar] [CrossRef] [Green Version]
  33. Zhijie, Z.; Qian, W.; Huadong, S.; Xuesong, J.; Qin, T.; Xiaoying, S. A novel sky region detection algorithm based on border points. Int. J. Signal Process. Image Process. Pattern Recognit. 2015, 8, 281–290. [Google Scholar]
  34. Wang, C.W.; Ding, J.J.; Chen, P.J. An efficient sky detection algorithm based on hybrid probability model. In Proceedings of the 2015 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA), Hong Kong, 16–19 December 2015; pp. 919–922. [Google Scholar]
  35. Irfanullah, K.H.; Sattar, Q.; Sadaqat-ur Rehman, A.A. An efficient approach for sky detection. IJCSI Int. J. Comput. Sci 2013, 10, 1694–1814. [Google Scholar]
  36. La Place, C.; Urooj, A.; Borji, A. Segmenting sky pixels in images: Analysis and comparison. In Proceedings of the 2019 IEEE Winter Conference on Applications of Computer Vision (WACV), Waikoloa Village, HI, USA, 7–11 January 2019; pp. 1734–1742. [Google Scholar]
  37. Sharma, P.; Schoemaker, M.; Pan, D. Automated Image Timestamp Inference Using Convolutional Neural Networks. Corpus ID: 159037027. 2016. Available online: https://www.semanticscholar.org/paper/Automated-Image-Timestamp-Inference-Using-Neural-Sharma/63f4e243aaefdea5cc14d453c12625242feeb2f1?sort=relevance&citationIntent=background (accessed on 30 November 2021).
  38. Hart, P.E.; Stork, D.G.; Duda, R.O. Pattern Classification; Wiley: Hoboken, NJ, USA, 2000. [Google Scholar]
  39. Fukunaga, K. Introduction to Statistical Pattern Recognition; Elsevier: Amsterdam, The Netherlands, 2013. [Google Scholar]
  40. Greenwood, P.E.; Nikulin, M.S. A Guide to Chi-Squared Testing; John Wiley & Sons: Hoboken, NJ, USA, 1996; Volume 280. [Google Scholar]
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.