Next Article in Journal
Cyclicity in EL–Hypergroups
Previous Article in Journal
A Fuzzy Classifier with Feature Selection Based on the Gravitational Search Algorithm
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

K-Hyperline Clustering-Based Color Image Segmentation Robust to Illumination Changes

1
School of Automation, Guangdong University of Technology, Guangzhou 510006, China
2
School of Physics and Electromechanical Engineering, Shaoguan University, Shaoguan 512026, China
*
Author to whom correspondence should be addressed.
Symmetry 2018, 10(11), 610; https://doi.org/10.3390/sym10110610
Submission received: 28 September 2018 / Revised: 26 October 2018 / Accepted: 4 November 2018 / Published: 7 November 2018

Abstract

:
Color image segmentation is very important in the field of image processing as it is commonly used for image semantic recognition, image searching, video surveillance or other applications. Although clustering algorithms have been successfully applied for image segmentation, conventional clustering algorithms such as K-means clustering algorithms are not sufficiently robust to illumination changes, which is common in real-world environments. Motivated by the observation that the RGB value distributions of the same color under different illuminations are located in an identical hyperline, we formulate color classification as a hyperline clustering problem. We then propose a K-hyperline clustering algorithm-based color image segmentation approach. Experiments on both synthetic and real images demonstrate the outstanding performance and robustness of the proposed algorithm as compared to existing clustering algorithms.

1. Introduction

Image segmentation, which partitions images into multiple regions with similar characteristics, is commonly used for image analysis and understanding. Simultaneously, it is also a fundamental step for further image understanding [1] and directly affects the accuracy of many computer vision applications [2], such as object detection [3], object tracking [4,5] and image retrieval [6]. In general, existing color image segmentation methods can be classified into five categories: thresholding, clustering, edge detection, region extraction and saliency detection [7]. These methods are mainly based on watershed transform or clustering procedures [8]. The former computes the spatial gradient of image luminance; however, the chromatic information is ignored [9]. The latter depends on the balance of the pixel color distribution [10].
The most popular clustering algorithm is K-means clustering, which is an exclusive clustering algorithm based on minimizing formal objective function [11]. Although K-means is widely used and studied, it is sensitive to outliers that lead to converge to local minima so that its accuracy is not perfect and may fail to represent the data [12]. In order to overcome the weaknesses of K-means clustering. Ahmed M N et al. [13] proposed Fuzzy C-Means (FCM), which allows a pixel belonging to a different cluster at multiple degrees of memberships and has been successfully applied to MR image segmentation. To further improve the noise immunity of FCM, Krinidis et al. [14] proposed the Fuzzy Local Information C-Means (FLICM) clustering algorithm, which incorporates the spatial and gray information in a fuzzy way. Although FLICM promotes noise immunity and is free of parameter selection, the calculation of the local information is time consuming and limits its application in high dimension data. To reduce the computational complexity and enhance the robustness to outlines, Lei et al. [15] introduced a Fast and Robust Fuzzy C-Means (FRFCM), which uses morphological reconstruction to smooth the image and employs a faster membership to calculate the distance between the pixels and cluster centers.
However, these local information-based clustering algorithms such as FLICM and FRFCM clustering are not robust to illumination changes, i.e., pixels with identical colors, but different illuminations may be segmented into different segments (sets of pixels), which limits their real-world applications. To handle the illumination changes in scenes, Li and Miao [16] employed a two-layer Gaussian mixture model to represent the different illumination backgrounds and segment the foreground image regions based on the learned model. Delibasis et al. [17] proposed a video segmentation robust to change in illumination based on different background models. Although these approaches achieve favorable performance on many video sequences, the supervised pre-learned background model limits their application in single image segmentation. Additionally, the distribution of color is not fully utilized.
Motivated by the observation that the RGB value distributions of the same kind of color, which is captured under different illuminations, are located in an identical hyperline, we employ hyperline clustering to perform unsupervised model-free color classification-based image segmentation such that it is robust to varying illuminations. Then, we propose a novel color classification-based image segmentation method using the multilayer K-hyperline clustering algorithm, which is capable of finding K partitions of the image data that converge to a global minimum. Finally, we conduct numerous experiments to verify the competitive result of the proposed approach.
The reminder of this paper is organized as follows. Section 2 presents the K-hyperline clustering method. Section 3 describes the proposed illumination-invariant color segmentation approach. Experiments on synthetic and color images from the Berkeley Segmentation Dataset (BSDS500) are carried out in Section 4. The conclusions and future work are presented in Section 5.

2. K-Hyperline Clustering

K-hyperline clustering is a special case of the general joint optimization problem of Sparse Component Analysis (SCA) [18]. The problem of SCA can be modeled as follows [19]:
Y = AX o r y = A x
where matrix Y = [ y 1 , , y T ] R m × T is an observed matrix, A = [ a 1 , , a n ] R m × n is an unknown full row rank basis matrix and X = [ x 1 , , x T ] R n × T is the matrix of sparse sources. When the sources X are sparse enough to satisfy the disjoint orthogonality condition approximately, i.e., x i · x j = 0 , the model (1) can be rewritten as:
y = A x = a 1 · x 1 + + a n · x n = a i · x i
Equation (2) can be further simplified as:
y 1 a i 1 = y 2 a i 2 = = y m a i m = x i
where a i = [ a i 1 , , a i m ] T . Since Equation (3) is a linear equation and all columns a 1 , , a m of A are the hyperline directions of the observed data y ( t ) . In other words, finding the basis matrix A can be considered as a hyperline clustering problem [20].
Given a set of observed data points y i i = 1 T , which respectively are located on K-hyperlines L ( l k ) , where l k is the directional vector of the corresponding hyperline and k = 1 , , K . K-Hyperline Clustering (K-HLC) aims to estimate K-hyperlines L ( l k ) by identifying l k , k = 1 , , K , as shown in Figure 1.
Mathematically, K-HLC can be considered as the following optimization problem [21]:
min l k , Ω k , k = 1 , , K J ( l k , Ω k , k = 1 , , K ) = min l k , Ω k , k = 1 , , K i = 1 T k = 1 K d 2 ( y i , l k ) × I i Ω k ,
where l k = [ l k 1 , , l k m ] R m is the directional vector of hyperline L ( l k ) ; the indicator function I i Ω k is given by:
I i Ω k = 1 , i Ω k 0 , i Ω k
Ω k denotes the k-th cluster set; the distance d ( y , l ) from y to L ( l k ) is:
d ( y , l ) = y , y y , l l , l ,
· , · denotes the inner product.
Recently, He et al. [21] proposed a K-Hyperline Clustering (K-HLC) algorithm, which is not only robust to outliers, but can also detect the number of hidden hyperlines in an unsupervised manner. Furthermore, K-hyperline clustering is capable of avoiding local minima with a high probability by incorporating Fast Multilayer Initialization (FMI).
The first step of FMI is to segment the initial direction matrix L ^ ( 0 ) = Y into M lower dimensional matrices L ^ 1 ( 0 ) , , L ^ M ( 0 ) , whose size are m × N M . Then, FMI constructs the initial direction matrix L ( 0 ) by selecting K possibly optimal columns from L ^ i ( 0 ) = [ l ^ i ( 0 ) ( 1 ) , , l ^ i ( 0 ) ( K + N M ) ] , i = 1 , , M . To select K optimal columns, all correlation coefficients c ( l i ( 0 ) ( k ) , y ( t ) ) = [ l ( 0 ) ] T [ y ( t ) ] are computed, and the sample y ( t ) is assigned to the k-th sets Ω ( l i ( 0 ) ( k ) ) if it satisfies k = arg max j = 1 , , K | c ( l i ( 0 ) ( j ) , y ( t ) | . Finally, the m × K matrix L ( 0 ) is constructed by extracting K columns corresponding to the greatest number of set Ω ( l i ( 0 ) ( k ) ) , k = 1 , , K + N M . Through FMI, a high-dimensional hyperline matrix can be rapidly scanned, and efficient initialization can be achieved.
After initialization using FMI, K-HLC is implemented in a manner similar to that of K-means clustering in terms of the two steps: cluster assignment and cluster centroid update. In the cluster assignment step, observed data y i i = 1 T are assigned to Ω k by choosing k such that:
k = arg min j = 1 , , K d 2 ( y i , l j )
In the second step, the cluster centroid is obtained by eigenvalue decomposition.

3. Color Image Segmentation Based on K-Hyperline Clustering

In this section, we illustrate the proposed color image segmentation method based on K-HLC in detail.
In the RGB color model, three primary colors (red, green and blue) are exploited together to produce an array of colors [22]. Thus, the RGB vector of a color pixel can be represented by:
y = [ r , g , b ] .
When the illumination changes, a simple and effective diagonal model, which maps colors under some reference illumination o to their corresponding color under another illumination u, is given by [23]:
R o G o B o = a 1 0 0 0 a 2 0 0 0 a 3 R u G u B u
Model (8) can be rewritten as:
y o = A y u
This model is similar to Model (1); thus, we employ K-HLC to solve the problem of illumination change.
To illustrate illumination changes, we generate a color grid with three primary colors—red ( [ 255 , 0 , 0 ] ) green ( [ 0 , 255 , 0 ] ) blue ( [ 0 , 0 , 255 ] )—and three secondary colors—yellow ( [ 255 , 255 , 0 ] ) magenta ( [ 255 , 0 , 255 ] ) cyan ( [ 0 , 255 , 255 ] )—under eight different illuminations, where we change their RGB values through reducing illumination using Photoshop software. The color grid image is shown in Figure 2a. Figure 2b shows that the pixel sequence { y i } i = 1 T of a color, for all illuminations on the color grid, is located on the same line, where:
y i = [ r i , g i , b i ] ,
i = 1 , , T and T is the total number of image pixels. In other words, the pixels of the original images and their corresponding varied illumination ones are located on the same line. In detail, it is obvious that six hyperlines (straight lines for 3D space) corresponding to six colors are located in six direction for the linear distribution of different illumination. The normalized directional vectors are [ 1 , 0 , 0 ] (red color), [ 0 , 1 , 0 ] (green color), [ 0 , 0 , 1 ] (blue color), [ 1 2 , 1 2 , 0 ] (yellow color), [ 1 2 , 0 , 1 2 ] (magenta color) and [ 0 , 1 2 , 1 2 ] (cyan color), respectively. Thus, the color classification can be formulated as the hyperline clustering problem. Moreover, to suppress the outliers and obtain good performance against spots, we normalize y i such that y i = 1 and initialize the hyperlines L ( l k ) using Fast Multilayer Initialization (FMI).
The proposed algorithm is summarized in Algorithm 1.
Algorithm 1: Color image segmentation based on K-HLC.
Input: Observed data y i i = 1 T and cluster number K.
1:
Initialize the direction of hyperline L ( 0 ) = [ l 1 ( 0 ) , , l K ( 0 ) ] using FMI and the number of iterations i t e r = 0 .
2:
Normalize y i such that y i = 1 , i = 1 , , T .
3:
while L ( i t e r ) L ( i t e r 1 ) < ε and i t e r < n u m _ i t e r * do
4:
 Assign data y i i = 1 T to K different clustering sets Ω k , k = 1 , , K . Data index i Ω k if and only if it satisfies k = a r g m i n i = 1 , , K d 2 ( y , l i ) .
5:
 Obtain a set of submatrices Y k = [ y 1 k , , y T ˜ k k ] , where T ˜ k data y 1 k , , y T ˜ k k are assigned to Ω k .
6:
 Compute the first eigenvector u 1 k of each matrix Y k · ( Y k ) T, k = 1 , , K .
7:
 Update direction vectors l k using u 1 k , k = 1 , , K .
8:
i t e r = i t e r + 1 .
9:
end while
Output: Clustering data Y k , k = 1 , , K .
* Usually, we can set n u m _ i t e r = 500 and ε = 0.05 .

4. Experimental Results and Discussion

We conducted numerical experiments to evaluate the performance of the proposed algorithm and performed comparisons with some existing methods. All experiments were tested in MATLAB R2015b, running on a Lenovo laptop with Intel I7 CPU 3.4 GHz and 8 GB RAM under Windows 7 Professional. First, we compared the proposed method with other algorithms in different color spaces. Second, we tested the synthetic images with different levels of illuminations. Third, we tested the method on images from the standard Berkeley Segmentation Dataset (BSDS500) [24]. The popular clustering segmentation methods: K-means, FCM, FILCM and FRFCM were employed in these for comparison, since the related codes are publicly accessible. The parameters of FILCM and FRFCM were selected following the suggestion of the authors. Besides, for a fair comparison, we set the same cluster number of all comparative approaches according to the ground truth without taking into account the spatial adjacency of pixels or the smoothness of segmentation boundaries. To evaluate the segmentation results quantitatively, we use Segmentation Accuracy (SA) as the performance measure, which is defined as follows [25]:
S A = Number of correctly classified pixels Total number of pixels

4.1. Results of Different Color Spaces

Besides widely used RGB color space, many color models, like HSL (Hue, Saturation, Lightness), HSV (Hue, Saturation, Value) and CIE-Lab, have been proposed to align with the human color vision system [25]. To demonstrate the superiority of the proposed method, we performed numerical experiments in different color spaces using Figure 2a. It is worth mentioning that we set the same RGB values of each grid. However, the marginal RGB value of each grid was different from the central one. This can be attributed to normalized function of MATLAB and is more challenging for a comparative algorithm. For the Lab color space, we utilized the MATLAB function “applycform” to convert the RGB value into the Lab color space. For HSL and HSV color spaces, to eliminate the effect of illuminations, we performed clustering procedures in the HS (Hue + Saturation) space. Since there are six colors in the color grid image, we set K = 6 in the experiment.
In Figure 3, we show the results of different clustering algorithms in Lab color spaces. The SA are reported below each segmentation result. From the reported SA, we can observe that FCM obtains the best segmentation result in the Lab color space. However, it tends to classify the lowest illumination as the same cluster, as shown in the last cluster of Figure 3b.
Figure 4 shows the results of the HS color space. Note that only the K-means and FCM are presented. The provided codes of both FLICM and FRFCM are designed in one or three dimensions. Although K-means performed better than FCM in the HS color space, there were still many spots in the results of K-means, as shown in the fifth cluster of Figure 4a.
Figure 5 shows that FRFCM obtained a better performance than FILCM due to its morphological reconstruction. However, it failed to cope with low illumination. Among all of them, the proposed K-HLC-based method gave the most accurate segmentation and obtained the highest SA value of one in the RGB color space. This can be attributed to the linear distribution of the color grid image, as shown in Figure 2.

4.2. Results of Synthetic Color Images

In the this test, we generated a set of synthetic color images, whose levels of illuminations varied from 50 % to 50%, as shown in Figure 6. For a fair comparison, We applied the proposed algorithm and other algorithms to both the original and normalized images. From Table 1, we can see that as the illumination changes, the SA values of normalized FCM clustering decreased much faster than others. This can be attributed to the performance of FCM completely depending on the location of the cluster center. It is worth noting that our method could correctly segment all 16 colors when illuminations varied from 40 % to 30%. However, other clustering methods failed to classify different color grids when they suffered illumination changes. As we can see, the proposed approach yielded the best results for all ten illumination levels.

4.3. Results of Real- World Color Images

In the third test, we selected five real-world images from BSDS500 (church, flower, rhinoceros, tiger and horses). Note that the results are only associated with image color information without considering the spatial connection of pixels, and the segment number K was set according to the ground truths. The obtained segmentation SAs are shown in Table 2. From the table, we can observe that our method provided much more satisfactory results than K-means and FCM clustering, particularly for the rhinoceros and horses images where the illumination of the foreground was homogeneous with that of background. Although FRFCM obtained a higher SA than ours for the flower image, other images results demonstrated the superiority of our method. Figure 7 shows the segmentation results for the selected example. Visual inspection demonstrates that our method out-performed both conventional and state-of-the-art clustering methods. We attribute this to the illumination invariance of our approach, which helped to segment a connecting foreground. For example, the results of the horses image obtained by the other algorithms were not able to segment the background as well as ours. This can be attributed to the illumination immunity of the proposed method.
To illustrate our approach further, we performed a scatter plot of three different regions of the church image, as shown in Figure 8. The wall, shadow and dome regions are represented by blue, red and black arrow lines, respectively. As we can observe, the location of the black cluster is closer to that of the red cluster; thus, the K-means and FCM tended to classify these two regions as the same color because of the distance between the clusters. In contrast, our approach could detect the hyperline of the blue and black cluster (denoted by purple line); hence, the proposed approach was able to identify the same color in different illuminations and obtained color coherent segmentation.

5. Conclusions

In this study, we formulate color image classification as a hyperline clustering problem and then develop a novel and effective K-hyperline clustering-based color image segmentation approach, which is robust to illumination changes. Both qualitative and quantitative analyses demonstrate that the proposed algorithm outperforms the conventional K-means and FCM and more recent FLICM and FRFCM clustering approaches in different color spaces. The limitation of the proposed approach is that K-HLC has a run time that is twice that of the K-means clustering approach. Therefore, our future work will focus on the development of a more efficient hyperline clustering algorithm and also its application to illumination-invariant visual tracking and object detection.

Author Contributions

Conceptualization, S.Y. Data curation, Y.X. Funding acquisition, Z.H. Methodology, S.Y. Project administration, Z.H. Software, P.L. and H.W. Supervision, Z.H. Validation, P.L., H.W. and Y.X. Writing, original draft, S.Y. Writing, review and editing, S.Y. and P.L.

Funding

This research was funded in part by the National Natural Science Foundation of China under Grants 61773127, 61773128 and 61727810, the Science and Technology Program of Shaoguan City of China under Grant No. SK201644, Guangdong Province Higher Vocational Colleges & Schools Pearl River Scholar Funded Scheme under Grant 2014 and the Natural Science Foundation of Guangdong Province of China under Grant 2016KQNCX156.

Acknowledgments

The authors would like to thank the editors and anonymous reviewers for their helpful comments and suggestions.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Chen, J.; Zheng, H.; Lin, X.; Wu, Y.; Su, M. A novel image segmentation method based on fast density clustering algorithm. Eng. Appl. Artif. Intell. 2018, 73, 92–110. [Google Scholar] [CrossRef]
  2. Gong, M.; Qian, Y.; Cheng, L. Integrated Foreground Segmentation and Boundary Matting for Live Videos. IEEE Trans. Image Process. 2015, 24, 1356–1370. [Google Scholar] [CrossRef] [PubMed]
  3. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 1137–1149. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  4. Čehovin, L.; Leonardis, A.; Kristan, M. Robust visual tracking using template anchors. In Proceedings of the 2016 IEEE Winter Conference on Applications of Computer Vision (WACV), Lake Placid, NY, USA, 7–10 March 2016; pp. 1–8. [Google Scholar] [CrossRef]
  5. Mazzon, R.; Cavallaro, A. Multi-camera tracking using a Multi-Goal Social Force Model. Neurocomputing 2013, 100, 41–50. [Google Scholar] [CrossRef]
  6. Jeong, J.; Won, I.; Yang, H.; Lee, B.; Jeong, D. Deformable Object Matching Algorithm Using Fast Agglomerative Binary Search Tree Clustering. Symmetry 2017, 9, 25. [Google Scholar] [CrossRef]
  7. Li, F.; Qin, J. Robust fuzzy local information and LpLp-norm distance-based image segmentation method. IET Image Process. 2017, 11, 217–226. [Google Scholar] [CrossRef]
  8. Yin, S.; Qian, Y.; Gong, M. Unsupervised Hierarchical Image Segmentation through Fuzzy Entropy Maximization. Pattern Recognit. 2017, 68, 245–259. [Google Scholar] [CrossRef]
  9. He, L.; Li, Y.; Zhang, X.; Chen, C.; Zhu, L.; Leng, C. Incremental Spectral Clustering via Fastfood Features and Its Application to Stream Image Segmentation. Symmetry 2018, 10, 272. [Google Scholar] [CrossRef]
  10. Moreno, R.; GranA, M.; Ramik, D.M.; Madani, K. Image segmentation on spherical coordinate representation of RGB colour space. Image Process. IET 2012, 6, 1275–1283. [Google Scholar] [CrossRef]
  11. Kanungo, T.; Mount, D.M.; Netanyahu, N.S.; Piatko, C.D.; Silverman, R.; Wu, A.Y. An Efficient k-Means Clustering Algorithm: Analysis and Implementation. IEEE Trans. Pattern Anal. Mach. Intell. 2002, 24, 881–892. [Google Scholar] [CrossRef]
  12. Nongmeikapam, K.; Kumar, W.K.; Singh, A.D. Fast and Automatically Adjustable GRBF Kernel Based Fuzzy C-Means for Cluster-wise Coloured Feature Extraction and Segmentation of MR Images. IET Image Process. 2018, 12, 513–524. [Google Scholar] [CrossRef]
  13. Ahmed, M.N.; Yamany, S.M.; Mohamed, N.; Farag, A.A.; Moriarty, T. A modified fuzzy C-means algorithm for bias field estimation and segmentation of MRI data. IEEE Trans. Med. Imag. 2002, 21, 193–199. [Google Scholar] [CrossRef] [PubMed]
  14. Krinidis, S.; Chatzis, V. A Robust Fuzzy Local Information C-Means Clustering Algorithm. IEEE Trans. Image Process. 2010, 19, 1328–1337. [Google Scholar] [CrossRef] [PubMed]
  15. Lei, T.; Jia, X.; Zhang, Y.; He, L.; Meng, H.; Nandi, A.K. Significantly Fast and Robust Fuzzy C-Means Clustering Algorithm Based on Morphological Reconstruction and Membership Filtering. IEEE Trans. Fuzzy Syst. 2018, 26, 3027–3041. [Google Scholar] [CrossRef]
  16. Li, J.; Miao, Z. Foreground segmentation for dynamic scenes with sudden illumination changes. Image Process. IET 2012, 6, 606–615. [Google Scholar] [CrossRef]
  17. Delibasis, K.K.; Goudas, T.; Maglogiannis, I. A novel robust approach for handling illumination changes in video segmentation. Eng. Appl. Artif. Intell. 2016, 49, 43–60. [Google Scholar] [CrossRef]
  18. Xie, K.; He, Z.; Cichocki, A.; Fang, X. Rate of Convergence of the FOCUSS Algorithm. IEEE Trans. Neural Netw. Learn. Syst. 2017, 28, 1276–1289. [Google Scholar] [CrossRef] [PubMed]
  19. Xie, K.; He, Z.; Cichocki, A. Convergence analysis of the FOCUSS algorithm. IEEE Trans. Neural Netw. Learn. Syst. 2017, 26, 601–613. [Google Scholar] [CrossRef] [PubMed]
  20. Bofill, P.; Zibulevsky, M. Underdetermined blind source separation using sparse representations. Signal Process. 2001, 81, 2353–2362. [Google Scholar] [CrossRef] [Green Version]
  21. He, Z.; Cichocki, A.; Li, Y.; Xie, S.; Sanei, S. K-hyperline clustering learning for sparse component analysis. Signal Process. 2009, 89, 1011–1022. [Google Scholar] [CrossRef] [Green Version]
  22. Swain, M.J.; Ballard, D.H. Color indexing. Int. J. Comput. Vision 1991, 7, 11–32. [Google Scholar] [CrossRef]
  23. Finlayson, G.D.; Drew, M.S.; Funt, B.V. Diagonal transforms suffice for color constancy. In Proceedings of the International Conference on Computer Vision, Berlin, Germany, 11–14 May 1993; pp. 164–171. [Google Scholar]
  24. Arbeláez, P.; Maire, M.; Fowlkes, C.; Malik, J. Contour detection and hierarchical image segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2011, 33, 898–916. [Google Scholar] [CrossRef] [PubMed]
  25. Smith, A.R. Color gamut transform pairs. ACM Siggraph Comput. Graph. 1978, 12, 12–19. [Google Scholar] [CrossRef]
Figure 1. Hyperline clustering, observed data points y i i = 1 T (blue points) and hidden hyperlines (red lines).
Figure 1. Hyperline clustering, observed data points y i i = 1 T (blue points) and hidden hyperlines (red lines).
Symmetry 10 00610 g001
Figure 2. (a) Color grid of six different colors under varying illuminations. (b) Scatter plot of the color grid in (a), where the colors of scatter points correspond to those of image pixels.
Figure 2. (a) Color grid of six different colors under varying illuminations. (b) Scatter plot of the color grid in (a), where the colors of scatter points correspond to those of image pixels.
Symmetry 10 00610 g002
Figure 3. Experimental results in the Lab color space. SA, Segmentation Accuracy; FLICM, Fuzzy Local Information C-Means; FRFCM, Fast and Robust Fuzzy C-Means.
Figure 3. Experimental results in the Lab color space. SA, Segmentation Accuracy; FLICM, Fuzzy Local Information C-Means; FRFCM, Fast and Robust Fuzzy C-Means.
Symmetry 10 00610 g003
Figure 4. Experimental results in the HS color space.
Figure 4. Experimental results in the HS color space.
Symmetry 10 00610 g004
Figure 5. Experimental results in the RGB color space.
Figure 5. Experimental results in the RGB color space.
Symmetry 10 00610 g005
Figure 6. Synthetic image, image with different level illuminations.
Figure 6. Synthetic image, image with different level illuminations.
Symmetry 10 00610 g006
Figure 7. Color image segmentation of the BSDS500 images (church, flower, rhinoceros, tiger and horses). The first row shows the original images. The second row shows the ground truths. The third row shows the results of K-means clustering. The fourth row shows the result of FCM clustering. The fifth row shows the results of FLICM. The sixth row shows the results of FRFCM. The seventh row shows the results of our method.
Figure 7. Color image segmentation of the BSDS500 images (church, flower, rhinoceros, tiger and horses). The first row shows the original images. The second row shows the ground truths. The third row shows the results of K-means clustering. The fourth row shows the result of FCM clustering. The fifth row shows the results of FLICM. The sixth row shows the results of FRFCM. The seventh row shows the results of our method.
Symmetry 10 00610 g007
Figure 8. Scatter plot of three different regions of the church image. The wall, shadow and dome regions are represented by blue, red and black arrow lines, respectively.
Figure 8. Scatter plot of three different regions of the church image. The wall, shadow and dome regions are represented by blue, red and black arrow lines, respectively.
Symmetry 10 00610 g008
Table 1. SA of comparative methods for synthetic images. The best result is shown in bold font.
Table 1. SA of comparative methods for synthetic images. The best result is shown in bold font.
Illumination Levels (%) 50 40 30 20 10 1020304050
K-means0.180.310.500.620.810.810.620.500.310.18
K-means (norm)0.250.370.560.680.810.810.810.500.310.25
FCM0.250.370.560.810.870.930.810.560.310.25
FCM (norm)0.060.060.120.120.180.180.120.120.120.06
FLICM0.320.420.560.650.810.810.680.620.480.35
FRFCM0.580.610.780.890.960.960.890.770.630.61
Our method 0.93 1 1 1 1 1 1 1 0.93 0.93
Table 2. Comparison of SA for the BSDS500 images. The best result is shown in bold font.
Table 2. Comparison of SA for the BSDS500 images. The best result is shown in bold font.
Images (Name)K-MeansFCMFLICMFRFCMOurs
Church0.92160.92460.82110.9730 0.9795
Flower0.83490.84570.7414 0.9648 0.9387
Rhinoceros0.55360.61050.62900.7266 0.9348
Tiger0.41800.65970.64610.8472 0.8597
Horses0.74250.79220.66430.8321 0.9257

Share and Cite

MDPI and ACS Style

Yang, S.; Li, P.; Wen, H.; Xie, Y.; He, Z. K-Hyperline Clustering-Based Color Image Segmentation Robust to Illumination Changes. Symmetry 2018, 10, 610. https://doi.org/10.3390/sym10110610

AMA Style

Yang S, Li P, Wen H, Xie Y, He Z. K-Hyperline Clustering-Based Color Image Segmentation Robust to Illumination Changes. Symmetry. 2018; 10(11):610. https://doi.org/10.3390/sym10110610

Chicago/Turabian Style

Yang, Senquan, Pu Li, HaoXiang Wen, Yuan Xie, and Zhaoshui He. 2018. "K-Hyperline Clustering-Based Color Image Segmentation Robust to Illumination Changes" Symmetry 10, no. 11: 610. https://doi.org/10.3390/sym10110610

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop