Next Article in Journal
Operational Surface Water Detection and Monitoring Using Radarsat 2
Previous Article in Journal
Error Analysis of Satellite Precipitation-Driven Modeling of Flood Events in Complex Alpine Terrain
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Performance Evaluation of Cluster Validity Indices (CVIs) on Multi/Hyperspectral Remote Sensing Datasets

1
Northeast Institute of Geography and Agroecology, Chinese Academy of Sciences, Changchun 130012, China
2
University of Chinese Academy of Sciences, Beijing 100049, China
3
Lancaster Environment Centre, Lancaster University, Lancaster LA1 4YQ, UK
4
Environmental Futures Research Institute, School of Environment, Griffith University, Brisbane, QLD 4111, Australia
*
Author to whom correspondence should be addressed.
Remote Sens. 2016, 8(4), 295; https://doi.org/10.3390/rs8040295
Submission received: 21 December 2015 / Revised: 26 February 2016 / Accepted: 21 March 2016 / Published: 30 March 2016

Abstract

:
The number of clusters (i.e., the number of classes) for unsupervised classification has been recognized as an important part of remote sensing image clustering analysis. The number of classes is usually determined by cluster validity indices (CVIs). Although many CVIs have been proposed, few studies have compared and evaluated their effectiveness on remote sensing datasets. In this paper, the performance of 16 representative and commonly-used CVIs was comprehensively tested by applying the fuzzy c-means (FCM) algorithm to cluster nine types of remote sensing datasets, including multispectral (QuickBird, Landsat TM, Landsat ETM+, FLC1, and GaoFen-1) and hyperspectral datasets (Hyperion, HYDICE, ROSIS, and AVIRIS). The preliminary experimental results showed that most CVIs, including the commonly used DBI (Davies-Bouldin index) and XBI (Xie-Beni index), were not suitable for remote sensing images (especially for hyperspectral images) due to significant between-cluster overlaps; the only effective index for both multispectral and hyperspectral data sets was the WSJ index (WSJI). Such important conclusions can serve as a guideline for future remote sensing image clustering applications.

Graphical Abstract

1. Introduction

Land use/cover data is crucial for diverse disciplines (e.g., ecology, geography, and climatology) since it serves as a basis for various “real world” applications [1,2,3]. Remote sensing technique have become the mainstream means to acquire land use/cover data, owing to its specific advantages, including synoptic views and cost-effectiveness [4,5]. Remote sensing image clustering, which utilizes only the statistical information inherent in the image without human interference, is one of the most widely used methods to produce land cover information [6,7]. It is also valued because of its high efficiency (i.e., it does not use training samples) [8].
The success of clustering (unsupervised classification) depends greatly on the proper determination of cluster number (i.e., the optimal number of classes) [9]: if the number of classes selected is less than the actual number, one or more separate classes would be merged into other classes; conversely, if larger, one or more homogeneous classes would be separated into different classes. The consequence is that the information contained in the raw data is incorrectly explored and used and the classification results will not be coincident with the “real” situation [10]. In this circumstance, the role of the cluster validity index (CVI), which is designed to detect the optimal cluster number for a given dataset, therefore, becomes critical [11].
Generally, a CVI is comprised of two indicators, namely compactness and separation. Compactness, which indicates the concentration of data points that belong to the same cluster, is usually measured by the distance between each data point and its cluster center [10]: the smaller the distance, the better the compactness of the cluster. Separation, which expresses the degree of isolation among clusters, is usually measured by the distance between cluster centroids: the larger the distance, the stronger the isolation of clusters [12]. Ideally, a dataset is partitioned with high compactness within each cluster and large separation between each pair of clusters. However, the two indicators are often mutually conflicting [13]; with increasing cluster number, the compactness becomes larger while the separation becomes smaller. Therefore, a good balance between the two indicators is required in the design of CVIs. To date, researchers from different disciplines have proposed a large number of CVIs for various types of applications.
In the remote sensing field, CVIs such as the Davies-Bouldin index (DBI) and the Xie-Beni index (XBI) have been widely used in image clustering applications. For example, DBI was employed to evaluate the fitness of candidate clustering by Bandyopadhyay and Maulik [9], and to guide satellite image clustering by Das et al. [14]; XBI was used to determine the optimal cluster number of IRS image by Maulik and Saha [15]; and was applied for multi-objective automatic image clustering [16,17]. However, in the absence of systematic and comprehensive evaluation of CVIs for remote sensing applications, CVIs are usually subjectively selected. This means that, without evaluation, they cannot necessarily be relied on. In fact, remote sensing data is well known for its complexity and uncertainty, with the specific characteristics as follows: (1) fuzzy and nonlinear class boundaries; (2) significant overlap among pixels from different classes (the overlap problem) [18]; and (3) high dimensionality and huge quantities of data. An appropriate CVI should, therefore, be designed taking account of these properties of remote sensing data.
To draw some general conclusions, although some efforts have been made to compare or evaluate the performance of CVIs in different environments [10,19,20,21], little attention has been paid to remote sensing data. Thus, the question remains as to how to select appropriate CVIs for remote sensing image clustering. Such a question can only be answered through an extensive evaluation of CVIs on various types of remote sensing data sets. However, to the best of our knowledge, few studies have addressed this issue. The objective of this paper is to fill that gap and identify one or several CVIs that are generally suitable for remote sensing datasets from a total of sixteen CVIs. The commonly used fuzzy c-means (FCM) and K-means algorithms were applied in this paper to cluster nine types of remote sensing datasets, including five types of multispectral and four types of hyperspectral images. This is of great significance since it can serve as a guideline for future remote sensing image clustering with diverse data types.
The remainder of this paper is organized as follows. In Section 2, the clustering problem and the FCM and K-means algorithms are briefly outlined; the sixteen CVIs evaluated in this paper are reviewed and detailed in Section 3; the experiments and results are provided in Section 4; the results are analyzed and discussed in Section 5; and conclusions are drawn in Section 6.

2. The Clustering Problem

In this section, we briefly review the clustering problem and the classical fuzzy c-means algorithm.

2.1. The Clustering Problem

Clustering is widely used in many fields to derive information on distributions and patterns in raw data [11]. It aims at partitioning a given data set into groups (clusters) according to a predefined criteria (usually the Euclidean distance) [20]. Let X = { x 1 , x 2 , , x N } be a possible given dataset (with N points), and K the number of clusters (i.e., patterns) of the data. The purpose of clustering is to evolve a partition matrix U ( X ) of the data to determine a partition C = { C 1 , C 2 , , C K } , in which the points in the same cluster are as close (i.e., have high similarity) as possible while those in different clusters are dispersed as far (i.e., have high dissimilarly) as possible. The partition matrix can be denoted as U = [ μ i j ] , 1 i K , 1 j N , where μ i j is the grade of membership of point x j to cluster C i ( i = 1 , , K ) .
Clustering can be performed in two forms: crisp and fuzzy. In crisp clustering, any one point of the given dataset belongs to only one class of clusters, that is μ i j = 1 if x j C i ; otherwise μ i j = 0 . In fuzzy clustering, a point may belong to several or all classes with a certain grade of membership. In this case, the partition matrix U ( X ) is represented as U = [ μ i j ] , where μ i j [ 0 , 1 ] . It should be noted that crisp clustering is a special version of fuzzy clustering in which the grade of membership of a point to a cluster is either 0 or 1. Once a fuzzy clustering structure is determined by a specific algorithm, each point of the given data will be assigned to the most likely cluster (i.e., with the largest grade of membership for that point). Through this process, the fuzzy clustering can be transformed into crisp clustering for real applications.

2.2. The Fuzzy C-Means Algorithm

The classical fuzzy c-means (FCM) algorithm proposed by Bezdek [22] has been successfully used in a wide domain of applications, such as agricultural engineering, image analysis, and target recognition, among others [20,23,24]. The objective of FCM is to evolve a set of cluster centers through minimizing the weighted within-cluster sum of squared error function J m , which is defined as:
J m = j = 1 N i = 1 K ( μ i j ) m x j z i 2 , 0 < j = 1 N μ i j < N , i { 1 , 2 , , K }
where Z = ( z 1 , z 2 , , z K ) is a group of cluster centers, z i R d ( d is the number of features included in each point). ||…|| is a Euclidean norm measuring the similarity between a point and the corresponding cluster center. The weighting exponent m controls the fuzziness of the grade of membership. The partition matrix μ i j and the cluster center set Z in the function J m can be calculated using the following equations:
μ i j = [ i = 1 K ( x j z i 2 x j z k 2 ) 1 / ( m 1 ) ] 1 , i { 1 , 2 , , K } , j { 1 , 2 , , N }
and
z i = j = 1 N ( μ i j ) m x j j = 1 N ( μ i j ) m , i { 1 , 2 , , K }
The FCM algorithm iteratively searches the fuzzy partition matrix and the cluster centers with a greedy searching strategy, until either no more changes are found in the cluster centers or the differences between two successive cluster centers fall below a predefined threshold. Normally, the FCM algorithm consists of the following steps:
Step 1: Determine the number of cluster K and the weighting exponent m , initialize the cluster centers Z ( z 1 , z 2 , , z K ) randomly, and define a threshold of iteration termination ε .
Step 2: Update the fuzzy partition matrix using Equation (2).
Step 3: Recalculate the cluster center set Z n e w using Equation (3).
Step 4: If Z n e w Z ε , stop the iteration and output the clustering result; otherwise, go to step 2.

2.3. The K-Means Algorithm

The K-means algorithm is one of the most commonly used methods for unsupervised image classification [2]. Similar to FCM, the objective of K-means is to determine a set of cluster centers through minimizing the clustering metric M , which is defined as
M = i = 1 K x j C i x j z i
where C i represents a cluster with z i as its cluster center.
A greedy searching strategy is also employed in K-means to search for the optimal set of cluster centers, until a predefined termination condition is met. The main steps of the algorithm are as follows:
Step 1: Determine the cluster number K and the maximum iteration number M a x _ i t e r to generate the initial cluster centers randomly.
Step 2: Assign pixel x j to cluster C i if x j z i < x j z k , k { 1 , 2 , , K } , and i k .
Step 3: Calculate new cluster center ( z i n e w ) for cluster C i as z i n e w = 1 N i x j C i x j , where N i denotes the number of pixels in cluster C i .
Step 4: If M a x _ i t e r is reached, terminate the cycle and output the clustering result; otherwise, go to Step 2.

3. Cluster Validity Indices (CVIs)

Broadly, current fuzzy CVIs (for fuzzy clustering) can be classified into two forms: one (called simple CVIs) only considers the fuzzy grades of membership to a class of the data (e.g., the partition coefficient), the other (called advanced CVIs) takes both fuzzy grads of membership and the geometrical properties (i.e., the structure) of the original data into account (e.g., the well-known XBI) [10]. In fact, crisp CVIs (for crisp clustering) which only consider the geometrical properties of the original data (e.g., the well-known DBI) are special versions of advanced CVIs, and can also be used in fuzzy image clustering analysis [12]. In this study, a total of 16 representative and commonly used CVIs of different forms were chosen for evaluation, including three simple CVIs, and thirteen advanced CVIs.
It is noteworthy that some CVIs (e.g., XBI) indicate the optimal cluster number of data by using the maximum value, while the others use the minimum value. For convenience, we subsequently denote the former (the larger, the better CVI) as CVI+, and the latter (the lower, the better CVI) as CVI.

3.1. Simple CVIs

(1) The partition coefficient (PC+) [25] evaluates the compactness by using the averaged strength of belongingness of data, and is defined as:
P C ( K ) = 1 N i = 1 K j = 1 N μ i j 2
(2) The partition entropy (PE) is formed based upon the logarithmic form of PC [22], and is defined as:
P E ( K ) = 1 N i = 1 K j = 1 N μ i j   log 2 ( u i j )
(3) The modification of PC (MPC+) [26] is designed to reduce the monotonic tendency of PC and PE. The index is defined as:
M P C ( K ) = 1 K K 1 ( 1 P C )

3.2. Advanced CVIs

(1) The Davies-Bouldin Index (DBI) [27] estimates the ratio of within-cluster compactness to between-cluster separation, which is defined as:
D B I ( K ) = 1 K i = 1 K max { S i + S k z i z k 2 } , i k ,
where S i = 1 N i x j C i x j z i 2 , N i denotes the number of data points in the i th cluster ( C i ).
(2) The Dunn Index (DI+) [28] evaluates a clustering by taking the minimum distance between-cluster as separation and the maximum distance between each pair of within-cluster points as compactness. The original index is defined as [28]:
D u n n ( K ) = min 1 p K ( min s + 1 q K 1 ( d i s ( C p , C q ) max 1 i K d i a ( C i ) ) ) ,
where d i s ( C p , C q ) refers to the distance between the p th and q th clusters, is calculated as d i s ( C p , C q ) = min x j C p , x l C q ( x j x l ) ; d i a ( C i ) denotes the maximum distance between any pair of within-cluster points, which is measured as d i a ( C i ) = max x j , x l C i ( x j x l ) .
(3) The Calinski-Harabasz Index (CHI+) [29] is a ratio-type index in which compactness is measured by the distance ( W K ) between each within-cluster point to its centroid, and separation is based on the distance ( B K ) between each centroid to the global centroid ( z ¯ ), i.e.,:
C H I ( K ) = B K K 1 / W K N K ,
where B K = i = 1 K N i z i z ¯ 2 , W K = i = 1 K x j C i x j z i 2 .
(4) The Fukuyama and Sugeno Index (FSI) [30] is designed to measure the discrepancy between fuzzy compactness and fuzzy separation, i.e.,:
F S I ( K ) = i = 1 K j = 1 N u i j m x j z i 2 i = 1 K j = 1 N u i j m z i z 2
(5) The Xie and Beni Index (XBI) [31] is also a ratio-type index, which measures the average within-cluster fuzzy compactness against the minimum between-cluster separation, i.e.,:
X B I ( K ) = i = 1 K j = 1 N μ i j 2 x j z i 2 N min i k { z i z k 2 }
(6) The Kwon Index (KI) [32] aims to overcome the shortcoming of XBI that decreases monotonically when the cluster number approaches the actual cluster number of data. Here, a penalty function was introduced to the numerator of XBI, i.e.,:
K I ( K ) = i = 1 K j = 1 N μ i j 2 x j z i 2 + 1 K i = 1 K z i z 2 min i k { z i z k 2 }
(7) The Tang Index (TI) [33] also introduced a similar penalty function to the numerator of XBI, i.e.:
T I ( K ) = i = 1 K j = 1 N μ i j 2 x j z i 2 + 1 K ( K 1 ) i = 1 K k = 1 k i K z i z k 2 min i k z i z k 2 + 1 / K
(8) The SC Index (SCI+) [34] measures the fuzzy compactness/separation ratio of clustering by using the difference between two functions, S C 1 and S C 2 , i.e.,:
S C I ( K ) = S C 1 ( K ) S C 2 ( K ) ,
where S C 1 (Equation (16)) evaluates the compactness/separation ratio by considering the grades of membership and the original data: the larger the S C 1 , the better the clustering:
S C 1 ( K ) = ( 1 K i = 1 K z i z ¯ ) i = 1 K ( j = 1 N μ i j m x j z i 2 / j = 1 N μ i j ) ,
while S C 2 (Equation (17)) measures the ratio by using the grades of membership only: the smaller the S C 2 , the better the clustering:
S C 2 ( K ) = i = 1 K 1 k = i + 1 K ( j = 1 N ( min ( μ i j , μ k j ) 2 ) / n j k ) ( j = 1 N max 1 i K μ i j 2 ) / ( j = 1 N max 1 i K μ i j )
where n j k = j = 1 N min ( μ i j , μ k j ) .
(9) The Compose Within and Between scattering Index (CWBI) [19] assesses the average compactness and separation of fuzzy clustering by using the sum of two functions, i.e.,:
C W B I ( K ) = α S c a t ( K ) + D i s ( K ) ,
where α is a weighing factor which equals D i s ( K max ) , the D i s ( K ) with the maximum cluster number; and S c a t ( K ) refers to the average scattering (i.e., compactness) for K clusters, which is defined as:
S c a t ( K ) = 1 K i = 1 K σ ( z i ) σ ( X ) ,
where x = ( x T x ) 1 / 2 ; σ ( X ) denotes the variance of data, which is defined as σ ( X ) = 1 N j = 1 N ( x j z ¯ ) 2 ; σ ( z i ) denotes the fuzzy variation of cluster i , which is defined as σ ( z i ) = 1 N j = 1 N μ i j ( x j z i ) 2 .
The smaller the value of S c a t ( K ) , the better the compactness of the clustering.
The distance function D i s ( K ) measuring the separation between clusters is defined as:
D i s ( K ) = D max D min i = 1 K ( k = 1 K z i z k ) 1 ,
where D max = max { z i z k } , D min = min { z i z k } , i , k {2, 3, … K }.
The smaller the value of D i s ( K ) , the better the separation of clusters.
(10) The WSJ Index (WSJI) [13], inspired by the CWBI, also uses a linear combination of averaged fuzzy compactness and separation to measure clustering, which is defined as:
W S J I ( K ) = S c a t ( K ) + S e p ( K ) S e p ( K max )
where S c a t ( K ) is given by Equation (19); S e p ( K ) denotes the between-cluster separation, which is defined as S e p ( K ) = D max 2 D min 2 i = 1 K ( k = 1 K z i z k 2 ) 1 , where D max = max { z i z k } , D min = min { z i z k } ; S e p ( K max ) refers to the S e p ( K ) with the maximum cluster number.
(11) The PBMF index (PBMFI+) [20] estimates within-cluster compactness and large separation between clusters of fuzzy clustering, i.e.,:
P B M F I ( K ) = max i k { z i z k } × j = 1 N μ j 1 x i z 1 K i = 1 K j = 1 N μ i j m x j z i .
(12) The SVF index (SVFI+) [35] emphasizes on low within-cluster variation (i.e., high compactness) and large separation between clusters, i.e.,:
S V F I ( K ) = i = 1 K min i k z i z k i = 1 K max x j C i μ i j m x j z i .
(13) The WL Index (WLI) [12] measures both within-cluster compactness and between-cluster separation of fuzzy clustering. Specifically, it takes both the minimum and the median distances between clusters as separation, which retains the clusters whose centroids are close to each other. The index is defined as:
W L I ( K ) = W L n 2 W L d
where W L n denotes the fuzzy compactness of clusters, which is defined as W L n = i = 1 K ( j = 1 N μ i j 2 x j z i 2 j = 1 N μ i j ) ; W L d refers to the separation between clusters, which is defined as W L d = 1 2 ( min i k { z i z k 2 } + m e d i a n i k { z i z k 2 } ) , where min i k { z i z k 2 } and m e d i a n { z i z k 2 } denote, respectively, the minimum distance and median distance between any pair of clusters.

4. Experiments and Results

In this section, the performance of the 16 CVIs introduced in Section 3 was evaluated using five types of multispectral, and four types of hyperspectral, remote sensing datasets (detailed below). For image clustering, the FCM and K-means algorithms were utilized here. The operational parameters in FCM were designated in line with previous studies [13]: threshold of iteration termination ε = e − 5, weighting exponent m = 2 , and the maximum iteration number M a x _ i t e r = 500 ; while the operational parameters in K-means as: the pixel change threshold = 0%, and the maximum iteration number M a x _ i t e r = 500 . For each of the images, the two algorithms were implemented with cluster number K = 2 , 3 , , 10 , respectively. To overcome the shortcoming of the two algorithms that often trap on local optima, depending on the initial solutions [36], each implementation of the clustering was repeated five times and the best clustering result (with the minimum value of J m (Equation (1)) or M (Equation (4)) was retained for CVIs evaluation.

4.1. Datasets

The five multispectral data sets include QuickBird [37], Landsat TM, Landsat ETM+, GaoFen-1 [38], and FLC1 [39]. Their true/false color maps, the corresponding ground reference maps and the spectral curves of land use/cover classes were shown in Figure 1. The four hyperspectral datasets include Hyperion [40], HYDICE [41], ROSIS [42] and AVIRIS [43]. Their false color maps, ground reference maps and spectral curves of land cover/use classes were presented in Figure 2. The basic information on the remote sensing datasets employed in our experiments was detailed in Table 1.

4.2. Results

The nine types of images were clustered by FCM and K-means algorithms respectively, and each clustering result was evaluated using the corresponding ground-truth data (Figure 1 and Figure 2). Table 2 shows the classification accuracies of the images achieved by the two algorithms. Similarly, both FCM and K-means generated good classification results, with the overall accuracy greater than 90% for seven images. However, considering the length limitation of the paper, the clustering results by K-means and the corresponding cluster validity result for each image were not presented in as much detail as those by FCM, but were summarized at the end of the results.
Table 3, Table 4, Table 5, Table 6, Table 7, Table 8, Table 9, Table 10 and Table 11 illustrated the variations of the 16 CVIs with the number of clusters ranging from two to 10 by FCM for each image. The optimal cluster numbers of each image are indicated by the CVIs, shown in bold font. The clustering results of multispectral and hyperspectral datasets by FCM, respectively, are illustrated in Figure 3 and Figure 4. Note that only four clustering results for each image are presented, including the optimal one (underlined), one or two close to the optimal, and those indicated by many CVIs (usually larger than 4) (bold). For example, Figure 3e–h illustrates the clustering results of Landsat TM image, in which the optimal clustering (Figure 3g) is underlined, the two near-optimal clustering results (Figure 3f,h) and the obviously-incorrect clustering indicated by many CVIs (Figure 3e) are also presented.
Figure 3a–d shows the clustering results of the simple QuickBird image with the cluster number K = 2 , 3 , 4 , 5 . The three ground truth classes (road, paddy field, and farmland) of the image were well identified with cluster number K = 3 (Figure 3b). As listed in Table 3, the majority of CVIs correctly indicated the actual cluster number of this simple image (except CHI, FSI, SCI, and SVFI).
Figure 3e-h illustrates the clustering results of the Landsat TM image with the cluster number K = 2 , 3 , 4 , 5 . Among them, the clustering with K = 4 succeeded in separating the four ground truth classes of the image (forest, farmland, water, and town) (Figure 3g). The clustering with K = 2 was obviously incorrect since three ground truth classes, i.e., forest, farmland, and town were merged into one class (Figure 3e). Unfortunately, as shown in Table 4 most indices (DBI, PC, PE, MPC, XBI, KI, TI, PBMFI, and WLI) underestimated the real situation, which preferred two as the cluster number of the image; whereas a clear overestimation was given by DI and SVFI; only five CVIs including CHI, FSI, SCI, CWBI, and WSJI provided the actual cluster number of the image.
Figure 3i–l portrays the clustering results of the Landsat ETM+ image with the cluster number K = 2 , 3 , 4 , 5 , respectively. The four ground truth classes of the image (marsh, forest, water, and farmland) were well separated with cluster number K = 4 (Figure 3k). However, similar to the Landsat TM experiment, most indices (PE, MPC, XBI, KI, TI, CWBI, PBMFI, and WLI) recommended two clusters as the optimal partitioning of the image (Table 5). CHI and WSJI were the only two indices that correctly indicated the cluster number of the image.
Figure 3m–p presents the clustering results of the GaoFen-1 image with the cluster number K = 2 , 4 , 5 , 6 . The five ground truth classes of the image, namely water1 (light colored), water2 (dark colored), grass, soil, and sand, were well distinguished with cluster number K = 5 (Figure 3o). This was correctly indicated by most CVIs including DBI, CHI, MPC, FSI, XBI, KI, TI, SCI, WSJI, and WLI (Table 6). For the rest of the CVIs that erroneously indicated the cluster number, most of them (DI, PC, PE, and PBMFI) suggested two.
Figure 3q–t demonstrates the clustering results of the FLC1 image with the cluster number K = 2 , 3 5 6 . The five ground truth classes of the image (soybeans, oats, corn, wheat, and red clover) were fairly well identified with cluster number K = 5 (Figure 3s). For the cases of clustering with K = 2 and K = 3 , obvious misclassifications were observed, with some separated classes being merged into one class (Figure 3q,r). As shown in Table 7, there were four CVIs (DI, CHI, FSI, and WSJI) that provided the actual cluster number ( K = 5 ) for the image. But five CVIs (DBI, PC, PE, PBMFI, and WLI) and five others (MPC, XBI, KI, TI, and CWBI) erroneously supported the clustering with cluster number K = 2 and K = 3 , respectively.
Figure 4a–d depicts the clustering results of Hyperion data with the cluster number K = 2 , 3 , 4 , 5 . The four ground truth classes (woodland, island interior, water, and floodplain grasses) were well classified with cluster number K = 4 (Figure 4c). Clustering results with other cluster numbers were obviously not satisfactory. For example, in the case of clustering with K = 2 , three (without water) of the four classes were wrongly merged into one class (Figure 4a). This was chosen by half the total CVIs, namely DBI, PC, PE, MPC, XBI, KI, TI, and CWBI (Table 8). In fact, all of the CVIs, except WSJI, failed to detect the actual cluster number of the image.
Figure 4e–h provides the clustering results of HYDICE data with the cluster number K = 3 , 4 , 5 , 6 , of which the clustering with K = 4 properly separated the four ground truth classes (roads, trees, trail, and grass) (Figure 4f). In the case of clustering with K = 3 , there were clear errors due to the incorrect merging of trees and roads (Figure 4e). However, it was still suggested by half of the CVIs, including DBI, MPC, XBI, KI, CWBI, TI, PBMFI, and WLI (Table 9). Similar to the experiment on Hyperion, WSJI was the only index returning the correct information about cluster number.
Figure 4i–l lists the clustering results of ROSIS data with the cluster number K = 3 , 4 , 5 , 6 . Among them, the clustering with K = 5 successfully classified the five ground truth classes (meadows, trees, asphalt, bricks, and shadows) (Figure 4k). For the case of K = 3 , trees and shadows were not distinguished (Figure 4i). However, this incorrect suggestion was also made by as many as half of the CVIs, including DBI, CHI, MPC, XBI, KI, TI, CWBI, and WLI (Table 10). Once again, only WSJI correctly indicated the actual cluster number of the image.
Figure 4m–p shows the clustering results of AVIRIS data with the cluster number K = 3 , 4 , 5 , 6 , in which the five ground truth classes (vineyard untrained, celery, fallow smooth, fallow plow, and stubble) were well classified with K = 5 (Figure 4o). However, no CVI was able to indicate the actual cluster number of the image (Table 11). Instead, most of them (DBI, DI, PC, MPC, XBI, KI, TI, CWBI, WSJI, and WLI) preferred four clusters for the image, which merged the classes of fallow smooth and fallow plow (Figure 4n).
Figure 5 illustrates the percentage of successes (correct guesses) achieved by the 16 CVIs. Table 12 summarizes the cluster validity results of the 16 CVIs by FCM on nine types of remote sensing image datasets, in which the actual cluster number of each image is listed in column K# while those indicated by CVIs are shown in other columns. From the table it can be seen that WSJI was the only index that correctly recognized the actual cluster numbers of all of the datasets (including multispectral and hyperspectral data), except for the AVIRIS image. Thus WSJI, was the most effective and stable index of all. CHI and FSI succeeded in multispectral datasets but failed in hyperspectral datasets. The DBI, DI, MPC, SCI, XBI, KI, TI, CWBI, and WLI indices were only effective for two multispectral images. CVIs including PC, PE, and PBMFI failed, generally, except for the simple QuickBird experiment. SVFI failed for all images.
Table 13 summarizes the cluster validity results of the 16 CVIs by K-means on the remote sensing images. As expected, the results are similar to those by FCM: WSJI performed the best for both multispectral and hyperspectral images, followed by CHI and FSI, both of which were effective for most multispectral images; the DBI, DI, MPC, SCI, XBI, KI, TI, CWBI, and PBMFI performed worse than the above-mentioned CVIs, with correct identification of cluster numbers for only one or two multispectral images; PC, PE, and SVFI behaved the worst since they failed for all images.

5. Discussion

In essence, a CVI is designed to measure the degree of compactness and/or separation of clusters. However, the two indicators are potentially in conflict because the number of clusters is generally positively associated with compactness, but negatively with separation. Thus, a balanced definition of compactness and separation is crucial to the designing of CVIs. Most CVIs measure compactness in a similar way, i.e., the distance between each data point and its cluster center. The major differences among CVIs, thus, lie in whether the indicator of separation is utilized and how it is defined. In fact, separation is not included in simple CVIs, whereas it is explicitly presented in all the advanced CVIs but with non-uniform definitions, e.g., the minimum or maximum distance between clusters. Beside issues of measuring compactness and separation, the performance of CVIs is also closely related to the nature of the experimental datasets. For remote sensing data, significant overlaps among clusters often exist [15]. This property of the datasets must, therefore, be considered by CVIs.
In our experiments, simple CVIs like PC, PE, and MPC were the worst performers. They underestimated the cluster numbers of most images (Table 10), consistent with previous studies [12,20]. This is mainly because they are built on the assumption that clusters are dispersed far away from each other, and the belonging (membership) of each point to its cluster is much larger than it is to other clusters. This assumption is, however, not necessarily valid in the context of the fuzzy property of remote sensing data. Thus, those CVIs without separation indicators are incapable of effectively handling such complex datasets.
Some advanced CVIs (including DBI, DI, XBI, KI, TI, PBMFI, SVFI, and WLI) usually performed better than simple CVIs, but were still far from satisfactory. This is mainly because these CVIs take the minimum distance (e.g., DBI and XBI) or the maximum distance (e.g., PBMFI) between clusters to measure separation and this results in a preference for clustering in which clusters are dispersed as far as possible [12]. However, clusters in remote sensing data are usually allocated closely. Those CVIs, in which separation exerts a great impact, therefore, underestimate the actual cluster numbers of the images (Table 10). It was found that some advanced CVIs (CHI, FSI, and SCI), in which the distance between each cluster center to the global center was taken as separation, worked better than the CVIs above. The separation measured with the distance from a single cluster to the global center, rather than the extreme (i.e., the minimum or maximum) distance between clusters, permits the existence of some clusters with close distances. However, they also failed in all experiments for hyperspectral images. The small distance between clusters in hyperspectral data weakened the role of separation in these CVIs, but enhanced the impact of compactness, thereby tending to overestimate the cluster number (Table 10).
Obviously, the bottleneck of most existing CVIs in handling large scale data (such as remote sensing data) lies in how to balance the two conflicting factors (compactness and separation) to correctly indicate the actual cluster number of the data. Fortunately, WSJI (the only index) strikes the right balance between the two factors through normalization [13], and its effectiveness is clearly verified in our experiments. This is especially suitable for handling complex remote sensing datasets: if the cluster number is underestimated, a large compactness emerges to penalize the clustering with too few clusters; conversely, if the cluster number is overestimated, a large separation appears to penalize the clustering with too many clusters; it is only when the actual cluster number is defined that both compactness and separation simultaneously become relatively small. The WSJI seemed to indicate a non-optimal clustering ( K = 4 ) only in the AVIRIS experiment, in which fallow smooth and fallow plow were merged into one cluster (Figure 4n). This is because the between-cluster distance of the two classes was small (having similar spectral characteristics (Figure 2l)), the value of separation increased significantly, so the WSJI did not indicate the optimal clustering ( K = 5 ). However, it was found that the two classes essentially belonged to the same land cover class (fallow), only differing in the surface roughness of the ground (smooth or plowed). In this sense, WSJI still detected the “real” cluster number of the image.
It should be noted that CVIs, a post-processing procedure applied to clustering, is based on the assumption that the structure of the dataset is well described by the cluster methods [35]. Otherwise, the evaluation would be meaningless [44]. Fortunately, the clustering results generated by the FCM algorithm worked fairly well in our experiments, as indicated by our land cover maps, although the cluster numbers were not very large. This is mainly due to the fact that it is difficult for FCM to handle complex images with a larger number of clusters [45,46]. In the future, it would be worthwhile to investigate more sophisticated methods (e.g., artificial intelligence-based algorithms) for image clustering. In addition to the 16 CVIs included in the experiments, some types of CVIs (such as graph-based validity measures [47]) that have recently appeared in the field of pattern recognition could be explored in further research to test their effectiveness in the area of remote sensing.
There was one dataset for each RS sensor in our experiment because of the length limitation of the paper. However, some remote sensing images, such as QuickBird versus Gaofen-1, and HYDICE versus AVIRIS, have great similarities in terms of number of bands, spatial resolution, and spectral wavelength. Two similar images can, thus, be confidently seen as two scenes of images acquired by effectively the same sensor. It is noted that the images employed in our experiments were acquired at different locations, which allows us to analyze the performance of CVIs on various types of landscapes (e.g., farmland, marsh and urban). Obviously, this might be helpful for the generalization of conclusions drawn in this paper.

6. Conclusions

In this paper, we evaluated the effectiveness of 16 representative CVIs on nine types of remote sensing datasets by utilizing the FCM algorithm for image clustering. From the experimental results, it was found that, due mainly to inappropriate definitions of separation, most existing CVIs were not suitable for remote sensing datasets (especially for hyperspectral data), which usually have significant overlaps between clusters. The only effective index was the WSJI, which indicated the real cluster numbers of the images (with either multispectral or hyperspectral data), because of its balanced combination between compactness and separation through normalization. This index, thus, deserves to be given first priority for future remote sensing classification applications. However, we do not claim that it would be effective in all applications because of the complexity of remote sensing datasets (e.g., a variety of formats, spatial scales, and spectral scales, among others). Furthermore, the selected algorithms and their operational parameters are very important for the performance of CVI, as they directly control the clustering quality (classification accuracy) of the image. In addition, the land use/cover classes covered by the images may also impact the cluster validity results. For example, WSJI failed in the AVIRIS experiment due to the very similar spectral characteristics of the two fallow classes (fallow smooth and fallow plow). As stated by Pal and Bezdek [48] “no matter how good your index is, there is a data set out there waiting to trick it (and you)”.

Acknowledgments

This work was supported by the National Natural Science Foundation of China (grant number: 41301465, 41271196), the National Major Program of China (grant number: 21-Y30B05-9001-13/15-2) and the “12th Five-Year Plan” Wetland and Blackland Specialised Databases Program of the Chinese Academy of Sciences (grant number: XXH12504-3-03). The authors would like to thank P. Gamba for providing the University of Pavia ROSIS data sets, and D. Landgrebe for providing the Washington DC HYDICE data sets. We thank the anonymous reviewers for their constructive comments.

Author Contributions

Huapeng Li conceived and designed the experiments; Huapeng Li and Xiaohui Ding performed the experiments; Huapeng Li, Shuqing Zhang and Ce Zhang analyzed the experimental results; Huapeng Li, Shuqing Zhang and Patricia Dale wrote the paper.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Kerr, J.T.; Ostrovsky, M. From space to species: Ecological applications for remote sensing. Trends Ecol. Evol. 2003, 18, 299–305. [Google Scholar] [CrossRef]
  2. Lu, D.; Weng, Q. A survey of image classification methods and techniques for improving classification performance. Int. J. Remote Sens. 2007, 28, 823–870. [Google Scholar] [CrossRef]
  3. Otukei, J.R.; Blaschke, T. Land cover change assessment using decision trees, support vector machines and maximum likelihood classification algorithms. Int. J. Appl. Earth Obs. Geoinf. 2010, 12, S27–S31. [Google Scholar] [CrossRef]
  4. Loveland, T.R.; Reed, B.C.; Brown, J.F.; Ohlen, D.O.; Zhu, Z.; Yang, L.; Merchant, J.W. Development of a global land cover characteristics database and IGBP DISCover from 1 km AVHRR data. Int. J. Remote Sens. 2000, 21, 1303–1330. [Google Scholar] [CrossRef]
  5. Friedl, M.A.; McIver, D.K.; Hodges, J.C.F.; Zhang, X.Y.; Muchoney, D.; Strahler, A.H.; Woodcock, C.E.; Gopal, S.; Schneider, A.; Cooper, A.; et al. Global land cover mapping from MODIS: Algorithms and early results. Remote Sens. Environ. 2002, 83, 287–302. [Google Scholar] [CrossRef]
  6. Xiao, X.M.; Boles, S.; Liu, J.Y.; Zhuang, D.F.; Liu, M.L. Characterization of forest types in Northeastern China, using multi-temporal SPOT-4 VEGETATION sensor data. Remote Sens. Environ. 2002, 82, 335–348. [Google Scholar] [CrossRef]
  7. Schmid, T.; Koch, M.; Gumuzzio, J.; Mather, P.M. A spectral library for a semi-arid wetland and its application to studies of wetland degradation using hyperspectral and multispectral data. Int. J. Remote Sens. 2004, 25, 2485–2496. [Google Scholar] [CrossRef]
  8. Duda, T.; Canty, M. Unsupervised classification of satellite imagery: Choosing a good algorithm. Int. J. Remote Sens. 2002, 23, 2193–2212. [Google Scholar] [CrossRef]
  9. Bandyopadhyay, S.; Maulik, U. Genetic clustering for automatic evolution of clusters and application to image classification. Pattern Recognit. 2002, 35, 1197–1208. [Google Scholar] [CrossRef]
  10. Wang, W.N.; Zhang, Y.J. On fuzzy cluster validity indices. Fuzzy Set Syst. 2007, 158, 2095–2117. [Google Scholar] [CrossRef]
  11. Halkidi, M.; Batistakis, Y.; Vazirgiannis, M. On clustering validation techniques. J. Intell. Inf. Syst. 2001, 17, 107–145. [Google Scholar] [CrossRef]
  12. Wu, C.H.; Ouyang, C.S.; Chen, L.W.; Lu, L.W. A new fuzzy clustering validity index with a median factor for centroid-based clustering. IEEE Trans. Fuzzy Syst. 2014, 23, 701–718. [Google Scholar] [CrossRef]
  13. Sun, H.J.; Wang, S.R.; Jiang, Q.S. FCM-based model selection algorithms for determining the number of clusters. Pattern Recognit. 2004, 37, 2027–2037. [Google Scholar] [CrossRef]
  14. Das, S.; Abraham, A.; Konar, A. Automatic clustering using an improved differential evolution algorithm. IEEE Trans. Syst Man Cybern. A 2008, 38, 218–237. [Google Scholar] [CrossRef]
  15. Maulik, U.; Saha, I. Automatic fuzzy clustering using modified differential evolution for image classification. IEEE Trans. Geosci. Remote Sens. 2010, 48, 3503–3510. [Google Scholar] [CrossRef]
  16. Zhong, Y.F.; Zhang, S.; Zhang, L.P. Automatic fuzzy clustering based on adaptive multi-objective differential evolution for remote sensing imagery. IEEE J. Sel. Top. Appl. Earth Observ. Remote Sens. 2013, 6, 2290–2301. [Google Scholar] [CrossRef]
  17. Ma, A.L.; Zhong, Y.F.; Zhang, L.P. Adaptive multiobjective memetic fuzzy clustering algorithm for remote sensing imagery. IEEE Trans. Geosci. Remote Sens. 2015, 53, 4202–4217. [Google Scholar] [CrossRef]
  18. Bandyopadhyay, S.; Pal, S.K. Pixel classification using variable string genetic algorithms with chromosome differentiation. IEEE Trans. Geosci. Remote Sens. 2011, 39, 303–308. [Google Scholar] [CrossRef]
  19. Rezaee, M.R.; Lelieveldt, B.P.F.; Reiber, J.H.C. A new cluster validity index for the fuzzy c-mean. Pattern Recognit. Lett. 1998, 19, 237–246. [Google Scholar] [CrossRef]
  20. Pakhira, M.K.; Bandyopadhyay, S.; Maulik, U. A study of some fuzzy cluster validity indices, genetic clustering and application to pixel classification. Fuzzy Sets Syst. 2005, 155, 191–214. [Google Scholar] [CrossRef]
  21. Arbelaitz, O.; Gurrutxaga, I.; Muguerza, J.; Perez, J.M.; Perona, I. An extensive comparative study of cluster validity indices. Pattern Recognit. 2013, 46, 243–256. [Google Scholar] [CrossRef]
  22. Bezdek, J.C. Partition Recognition with Fuzzy Objective Function Algorithms; Plenum Press: NewYork, NY, USA, 1981. [Google Scholar]
  23. Bezdek, J.C. Partition Structures: A Tutorial in the Analysis of Fuzzy Information; CRC Press: Boca Raton, FL, USA, 1987. [Google Scholar]
  24. Cai, W.L.; Chen, S.C.; Zhang, D.Q. Fast and robust fuzzy c-means clustering algorithms incorporating local information for image segmentation. Pattern Recognit. 2007, 40, 825–838. [Google Scholar] [CrossRef]
  25. Bezdek, J.C. Cluster validity with fuzzy sets. J. Cybern. 1974, 3, 58–72. [Google Scholar] [CrossRef]
  26. Dave, R.N. Validating fuzzy partition obtained throughc-shells clustering. Pattern Recognit. Lett. 1996, 17, 613–623. [Google Scholar] [CrossRef]
  27. Davies, D.L.; Bouldin, D.W. A clustering separation measure. IEEE Trans. Pattern Anal. Mach. Intell. 1979, 1, 224–227. [Google Scholar] [CrossRef] [PubMed]
  28. Dunn, J.C. A fuzzy relative of the ISODATA process and its use in detecting compact well-separated clusters. J. Cybern. 1973, 3, 32–57. [Google Scholar] [CrossRef]
  29. Calinski, T.; Harabasz, J. A dendrite method for cluster analysis. Commun Stat. Theory Methods 1974, 3, 1–27. [Google Scholar] [CrossRef]
  30. Fukuyama, Y.; Sugeno, M. A new method of choosing the number of clusters for fuzzy c-means method. In Proceedings of the 5th Fuzzy System Symposium, Kobe, Japan, 2–3 June 1989.
  31. Xie, X.L.; Beni, G. A validity measure for fuzzy clustering. IEEE Trans. Pattern Anal. 1991, 13, 841–847. [Google Scholar] [CrossRef]
  32. Kwon, S.H. Cluster validity index for fuzzy clustering. Electron. Lett. 1998, 34, 2176–2177. [Google Scholar] [CrossRef]
  33. Tang, Y.G.; Sun, F.C.; Sun, Z.Q. Improved validation index for fuzzy clustering. In Proceedings of the American Control Conference, Portland, OR, USA, 8–10 June 2005.
  34. Zahid, N.; Limouri, N.; Essaid, A. A new cluster-validity for fuzzy clustering. Pattern Recognit. 1999, 32, 1089–1097. [Google Scholar] [CrossRef]
  35. Zalik, K.R.; Zalik, B. Validity index for clusters of different sizes and densities. Pattern Recognit. Lett. 2011, 32, 221–234. [Google Scholar] [CrossRef]
  36. Jain, A.K. Data clustering: 50 years beyond K-means. Pattern Recognit. Lett. 2010, 31, 651–666. [Google Scholar] [CrossRef]
  37. Wang, L.; Sousa, W.P.; Gong, P.; Biging, G.S. Comparison of IKONOS and QuickBird images for mapping mangrove species on the Caribbean coast of Panama. Remote Sens. Environ. 2004, 91, 432–440. [Google Scholar] [CrossRef]
  38. Li, J.; Chen, X.L.; Tian, L.Q.; Huang, J.; Feng, L. Improved capabilities of the Chinese high-resolution remote sensing satellite GF-1 for monitoring suspended particulate matter (SPM) in inland waters: Radiometric and spatial considerations. ISPRS J. Photogramm. Remote Sens. 2015, 106, 145–156. [Google Scholar] [CrossRef]
  39. Tadjudin, S.; Landgrebe, D.A. Robust parameter estimation for mixture model. IEEE Trans. Geosci. Remote Sens. 2000, 38, 439–445. [Google Scholar] [CrossRef]
  40. Ham, J.; Chen, Y.C.; Crawford, M.M.; Ghosh, J. Investigation of the random forest framework for classification of hyperspectral data. IEEE Trans. Geosci. Remote Sens. 2005, 43, 492–501. [Google Scholar] [CrossRef]
  41. Yin, J.H.; Wang, Y.F.; Hu, J.K. A new dimensionality reduction algorithm for hyperspectral image using evolutionary strategy. IEEE Trans. Ind. Inform. 2012, 8, 935–943. [Google Scholar] [CrossRef]
  42. Li, J.; Bioucas-Dias, J.M.; Plaza, A. Spectral-spatial hyperspectral image segmentation using subspace multinomial logistic regression and markov random fields. IEEE Trans. Geosci. Remote Sens. 2012, 50, 809–823. [Google Scholar] [CrossRef]
  43. Zhou, Y.C.; Peng, J.T.; Chen, C.L.P. Extreme learning machine with composite kernels for hyperspectral image classification. IEEE J. Selected Topics Appl. Earth Observ. Remote Sens. 2015, 8, 2351–2360. [Google Scholar] [CrossRef]
  44. Kim, D.W.; Lee, K.H.; Lee, D.H. On cluster validity index for estimation of the optimal number of fuzzy clusters. Pattern Recognit. 2004, 37, 2009–2025. [Google Scholar] [CrossRef]
  45. Maulik, U.; Bandyopadhyay, S. Fuzzy partitioning using a real-coded variable-length genetic algorithm for pixel classification. IEEE Trans. Geosci. Remote Sens. 2003, 41, 1075–1081. [Google Scholar] [CrossRef]
  46. Paoli, A.; Melgani, F.; Pasolli, E. Clustering of hyperspectral images based on multiobjective particle swarm optimization. IEEE Trans. Geosci. Remote Sens. 2009, 47, 4175–4188. [Google Scholar] [CrossRef]
  47. Ilc, N. Modified Dunn’s cluster validity index based on graph theory. Electr. Rev. 2012, 2, 126–131. [Google Scholar]
  48. Pal, N.R.; Bezdek, J.C. Correction to “on cluster validity for the fuzzy c-means model”. IEEE Trans. Fuzzy Syst. 1997, 5, 152–153. [Google Scholar] [CrossRef]
Figure 1. The multispectral images. (ac) the true/false color map, the ground reference map and the corresponding spectral curves of ground truth classes of QuickBird datasets; (df) the corresponding maps of Landsat TM datasets; (gi) the corresponding maps of Landsat ETM+ datasets; (jl) the corresponding maps of GaoFen-1 datasets; (mo) the corresponding maps of FLC1 datasets. (a) True color map; (b) false color map (7, 5, 3); (c) false color map (7, 5, 3); (d) true color map; and (e) false color map (bands 12, 9, and 1).
Figure 1. The multispectral images. (ac) the true/false color map, the ground reference map and the corresponding spectral curves of ground truth classes of QuickBird datasets; (df) the corresponding maps of Landsat TM datasets; (gi) the corresponding maps of Landsat ETM+ datasets; (jl) the corresponding maps of GaoFen-1 datasets; (mo) the corresponding maps of FLC1 datasets. (a) True color map; (b) false color map (7, 5, 3); (c) false color map (7, 5, 3); (d) true color map; and (e) false color map (bands 12, 9, and 1).
Remotesensing 08 00295 g001aRemotesensing 08 00295 g001b
Figure 2. The hyperspectral images. (ac) the false color (FC) map, the corresponding ground reference map and the corresponding spectral curves of ground truth classes of Hyperion data sets; (df) the corresponding maps of HYDICE datasets; (gi) the corresponding maps of ROSIS datasets; (jl) the corresponding maps of AVIRIS datasets. (a) FC map (bands 93, 60, 10); (b) FC map (bands 120, 90, 10); (c) FC map (bands 90, 60, 10); and (d) FC map (bands 111, 90, 12).
Figure 2. The hyperspectral images. (ac) the false color (FC) map, the corresponding ground reference map and the corresponding spectral curves of ground truth classes of Hyperion data sets; (df) the corresponding maps of HYDICE datasets; (gi) the corresponding maps of ROSIS datasets; (jl) the corresponding maps of AVIRIS datasets. (a) FC map (bands 93, 60, 10); (b) FC map (bands 120, 90, 10); (c) FC map (bands 90, 60, 10); and (d) FC map (bands 111, 90, 12).
Remotesensing 08 00295 g002
Figure 3. Clustering results of the multispectral images (each color represents a cluster). (a) QuickBird, K = 2; (b) QuickBird, K = 3; (c) QuickBird, K = 4; (d) QuickBird, K = 5; (e) Landsat TM, K = 2; (f) Landsat TM, K = 3; (g) Landsat TM, K = 4; (h) Landsat TM, K = 5; (i) Landsat ETM+, K = 2; (j) Landsat ETM+, K = 3; (k) Landsat ETM+, K = 4; (l) Landsat ETM+, K = 5; (m) GaoFen-1, K = 2; (n) GaoFen-1, K = 4; (o) GaoFen-1, K = 5; (p) GaoFen-1, K = 6; (q) FLC1, K = 2; (r) FLC1, K = 3; (s) FLC1, K = 5; and (t) FLC1, K = 6.
Figure 3. Clustering results of the multispectral images (each color represents a cluster). (a) QuickBird, K = 2; (b) QuickBird, K = 3; (c) QuickBird, K = 4; (d) QuickBird, K = 5; (e) Landsat TM, K = 2; (f) Landsat TM, K = 3; (g) Landsat TM, K = 4; (h) Landsat TM, K = 5; (i) Landsat ETM+, K = 2; (j) Landsat ETM+, K = 3; (k) Landsat ETM+, K = 4; (l) Landsat ETM+, K = 5; (m) GaoFen-1, K = 2; (n) GaoFen-1, K = 4; (o) GaoFen-1, K = 5; (p) GaoFen-1, K = 6; (q) FLC1, K = 2; (r) FLC1, K = 3; (s) FLC1, K = 5; and (t) FLC1, K = 6.
Remotesensing 08 00295 g003
Figure 4. Clustering results of the hyperspectral images by FCM (each color represents a cluster). (a) Hyperion, K = 2; (b) Hyperion, K = 3; (c) Hyperion, K = 4; (d) Hyperion, K = 5; (e) HYDICE, K = 3; (f) HYDICE, K = 4; (g) HYDICE, K = 5; (h) HYDICE, K = 6; (i) ROSIS, K = 3; (j) ROSIS, K = 4; (k) ROSIS, K = 5; (l) ROSIS, K = 6; (m) AVIRIS, K = 3; (n) AVIRIS, K = 4; (o) AVIRIS, K = 5; and (p) AVIRIS, K = 6.
Figure 4. Clustering results of the hyperspectral images by FCM (each color represents a cluster). (a) Hyperion, K = 2; (b) Hyperion, K = 3; (c) Hyperion, K = 4; (d) Hyperion, K = 5; (e) HYDICE, K = 3; (f) HYDICE, K = 4; (g) HYDICE, K = 5; (h) HYDICE, K = 6; (i) ROSIS, K = 3; (j) ROSIS, K = 4; (k) ROSIS, K = 5; (l) ROSIS, K = 6; (m) AVIRIS, K = 3; (n) AVIRIS, K = 4; (o) AVIRIS, K = 5; and (p) AVIRIS, K = 6.
Remotesensing 08 00295 g004
Figure 5. The overall performance of CVIs by applying FCM algorithm to cluster nine types of remote sensing datasets.
Figure 5. The overall performance of CVIs by applying FCM algorithm to cluster nine types of remote sensing datasets.
Remotesensing 08 00295 g005
Table 1. Basic information of the remote sensing data sets.
Table 1. Basic information of the remote sensing data sets.
DSYLRBWSGT
QuickBirdMulti-spectral camera2005Yalvhe farm, China2.440.45–0.90100 × 100Road, paddy field, and farmland
Landsat TMThematic mapper2005JingYuetan reservoir, China3060.45–2.35296 × 295Forest, farmland, water, and town
Landsat ETM+Enhanced thematic mapper2001Zhalong reserve, China3060.45–2.35150 × 139Marsh, forest, water, and farmland
Gaofen-1Wide filed imager2015Sanjiang Plain, China1640.45–0.89200 × 200Water1, water2, grass, soil, and sand
FLC1M7 scanner1966Tippecanoe County, US30120.40–1.0084 × 183Soybeans, oats, corn, wheat and red clover
HyperionHyperion2001Okavango Delta, Botswana301450.40–2.50126 × 146Woodland, island interior, water and floodplain grasses
HYDICEHYDICE1995Washington DC, US21910.40–2.40126 × 82Roads, trees, trail and grass
ROSISROSIS2001University of Pavia, Italy1.31030.43–0.86125 × 148Meadows, trees, asphalt, bricks and shadows
AVIRISAVIRIS1998Salinas Valley, USA3.72040.41–2.45117 × 143Vineyard untrained, celery, fallow smooth, fallow plow and stubble
Note: D, datasets; S, sensor; Y, year; L, location; R, resolution (m); B, number of bands; W, spectral wavelength (µm); S, size of image (pixel by pixels); GT, ground truth classes.
Table 2. Classification accuracies of the remote sensing images acquired by FCM and K-means algorithms.
Table 2. Classification accuracies of the remote sensing images acquired by FCM and K-means algorithms.
DatasetsK#Overall Accuracies (%)Kappa Coefficient
FCMK-MeansFCMK-Means
QuickBird396.0696.100.93540.9361
Landsat TM495.7895.270.94330.9363
Landsat ETM+494.4196.300.92530.9565
Gaofen-1598.3498.790.97910.9848
FLC1583.1084.480.78470.8016
Hyperion487.0986.790.82600.8219
HYDICE494.8896.000.92380.9403
ROSIS593.8593.250.91290.9044
AVIRIS599.6399.630.99460.9946
Table 3. Variations of the 16 CVIs with cluster numbers ranging from 2 to 10 for the QuickBird image.
Table 3. Variations of the 16 CVIs with cluster numbers ranging from 2 to 10 for the QuickBird image.
CVIsCluster Number
23 *45678910
PC+0.7540.8000.7280.6990.6580.6230.6260.5930.579
PE0.5650.5320.7490.8661.0081.1341.1501.2721.341
MPC+0.5090.7000.6370.6240.5900.5600.5720.5420.533
DBI0.8220.5590.6920.7500.7790.8450.8020.8820.915
DI+(e-3)2.4863.5912.5392.5392.6142.6143.4393.4393.439
CHI+(e4)1.1422.0412.0262.0301.9081.7792.1161.9981.971
FSI(e7)−0.535−6.644−7.247−7.331−7.093−6.902−7.440−7.260−7.092
XBI0.1600.1030.2180.2360.2310.2870.2370.3250.277
KI(e3)1.6011.0272.1842.3702.3122.8832.3823.2662.790
TI(e3)1.6011.0292.1892.3762.3202.8942.3953.2842.808
SCI+0.4772.4772.5162.7522.8372.5543.5463.4563.349
CWBI(e-2)4.0762.8264.2844.9645.3996.9026.5928.6288.499
WSJI0.3650.1710.2560.3390.3990.6480.6181.0601.023
PBMFI+(e3)1.5883.2140.6350.3360.0680.0600.0220.0340.013
SVFI+1.4222.4152.6212.9103.0633.2523.5173.6993.885
WLI0.3390.2520.3250.4180.4260.4530.3460.3740.388
Note: * denotes the actual cluster number of the image; figures in bold face denote the optimal cluster numbers of the image identified by the CVIs; the data in the brackets of the first column is a multiplying factor (e.g., e-3 followed DI+) of the corresponding line.
Table 4. Variations of the 16 CVIs with cluster numbers ranging from 2 to 10 for the Landsat TM image.
Table 4. Variations of the 16 CVIs with cluster numbers ranging from 2 to 10 for the Landsat TM image.
CVIsCluster Number
234 *5678910
PC+0.9220.7960.7940.6890.6570.6100.5870.5710.472
PE0.2040.5210.5870.8560.9981.1681.2811.3541.593
MPC+0.8440.6950.7260.6120.5890.5450.5280.5170.413
DBI0.2820.7640.6010.9990.9131.0551.0641.1851.522
DI+(e-3)8.007.5487.7838.0428.4988.8938.8938.8938.893
CHI+(e5)2.8832.6083.2612.9142.8312.5682.3932.3482.053
FSI(e7)−6.762−7.850−8.743−8.423−8.342−8.132−8.027−7.945−5.888
XBI0.0440.1770.0920.5990.4500.6260.5640.6612.025
KI(e4)0.3341.3350.6934.5233.4024.7314.2594.99615.300
TI(e4)0.3341.3350.6934.5153.3974.7214.2514.98515.188
SCI+3.4633.3193.8763.6093.4072.9163.2042.3623.439
CWBI(e-2)0.1510.1420.1140.2910.2990.4030.4250.4570.762
WSJI0.1670.0940.0570.1540.1590.2740.3070.3721.015
PBMFI+(e3)1.0140.2960.0180.0130.0060.0060.0010.0010.001
SVFI+1.9222.2872.9823.3673.8164.1064.3494.2243.137
WLI0.0790.1310.1720.2470.3270.3610.4630.3950.293
Note: * denotes the actual cluster number of the image.
Table 5. Variations of the 16 CVIs with cluster numbers ranging from 2 to 10 for the ETM+ image.
Table 5. Variations of the 16 CVIs with cluster numbers ranging from 2 to 10 for the ETM+ image.
CVIsCluster Number
234 *5678910
PC+0.4040.8120.7600.7250.7030.6630.6280.6170.593
PE0.2840.5030.6730.8020.8771.0121.1441.2041.300
MPC+0.7750.7170.6800.6560.6440.6070.5750.5690.548
DBI0.4040.6030.6740.7310.7450.8610.9520.9371.055
DI+(e-3)5.8037.5958.2568.8898.8890.1040.1070.1070.114
CHI+(e5)0.8180.9090.9290.9290.8960.8930.8470.8470.822
FSI(e8)−0.409−0.483−0.477−0.461−0.461−0.439−0.426−0.422−0.413
XBI0.0470.0900.1170.1690.1830.2020.2340.2160.293
KI(e4)0.1000.1880.2440.3520.3820.4210.4890.4520.613
TI(e4)0.0990.1890.2440.3530.3830.4220.4900.4540.615
SCI+2.4633.2593.3883.7234.7155.1374.9224.7794.884
CWBI0.0540.0590.0760.1040.1120.1330.1640.1740.226
WSJI0.1620.3510.1260.2140.2480.3480.5150.6011.012
PBMFI+(e3)1.0580.1560.0850.0880.0040.0050.0020.0140.003
SVFI+2.1732.4133.0270.7603.1553.3393.5213.6073.648
WLI0.0930.1740.3070.2650.2440.2290.2590.2080.279
Note: * denotes the actual cluster number of the image.
Table 6. Variations of the 16 CVIs with cluster numbers ranging from 2 to 10 for the GaoFen-1 image.
Table 6. Variations of the 16 CVIs with cluster numbers ranging from 2 to 10 for the GaoFen-1 image.
CVIsCluster Number
2345 *678910
PC+0.8500.7510.7640.7790.7350.7100.6880.6690.645
PE0.3740.6430.6630.6581.1830.9000.9781.0561.140
MPC+0.7000.6270.6200.7240.6820.6620.6430.6280.606
DBI0.5870.8320.7670.5360.7010.7400.7890.8030.896
DI+(e-3)4.7151.4781.4702.2982.3482.6883.0282.8602.965
CHI+(e5)0.7640.6560.6811.3141.1851.2621.2701.2571.197
FSI(e9)−0.662−1.038−1.253−1.603−1.571−1.539−1.519−1.490−1.462
XBI0.1000.1350.1360.0770.1980.1580.1400.1540.285
KI(e4)0.3990.5410.5460.3090.7910.6340.5600.6171.141
TI(e4)0.3990.5410.5460.3090.7920.6350.5610.6181.141
SCI+1.0311.1561.3644.6583.8994.4684.4704.3333.992
CWBI(e-3)0.1470.1370.1440.1390.2400.2640.2680.3110.445
WSJI0.2260.9340.1280.1140.2820.3480.3620.4871.015
PBMFI+(e4)1.0490.2430.0750.0530.0630.0190.0110.0400.056
SVFI+2.3152.6103.0273.7073.7394.1934.0094.2454.165
WLI0.2010.3420.3070.1540.1740.2020.1960.2080.226
Note: * denotes the actual cluster number of the image.
Table 7. Variations of the 16 CVIs with cluster numbers ranging from 2 to 10 for the FLC1 image.
Table 7. Variations of the 16 CVIs with cluster numbers ranging from 2 to 10 for the FLC1 image.
CVIsCluster Number
2345 *678910
PC+0.7600.6800.5840.6020.5550.4970.4510.4290.398
PE0.5490.8341.1391.1601.3431.5561.7241.8311.974
MPC+0.5190.5200.4460.5030.4660.4140.3720.3580.331
DBI0.8870.9091.0520.8960.9371.0081.4321.4101.475
DI+(e-2)0.8061.3301.0481.6131.3651.4951.2591.2591.542
CHI+(e4)1.2741.5371.2731.6161.5211.3541.2491.1941.105
FSI(e6)−0.666−5.029−5.479−8.809−8.573−7.979−7.511−7.200−6.794
XBI0.2060.1860.3800.2240.2680.3340.6360.6160.576
KI(e4)0.2990.2700.5520.3250.3900.4850.9240.8960.837
TI(e4)0.2990.2700.5520.3250.3900.4850.9240.8960.838
SCI+0.3070.6240.3830.9131.0890.6720.6390.8150.681
CWBI0.1260.0980.1230.1090.1280.1540.2210.2300.238
WSJI0.4100.5980.2940.2410.3050.4250.8770.9681.049
PBMFI+186.17884.91527.8915.07616.6385.9695.3742.1400.932
SVFI+1.1851.8742.3893.0403.4243.6803.4393.6793.803
WLI0.4150.5350.7080.5230.5840.7097.5910.7270.777
Note: * denotes the actual cluster number of the image.
Table 8. Variations of the 16 CVIs with cluster numbers ranging from 2 to 10 for the Hyperion image.
Table 8. Variations of the 16 CVIs with cluster numbers ranging from 2 to 10 for the Hyperion image.
CVIsCluster Number
234 *5678910
PC+0.8670.7590.6820.6580.5960.5680.5300.4940.476
PE0.3370.6260.8690.9731.1761.2931.4351.5771.666
MPC+0.7350.6380.5760.5730.5150.4960.4630.4300.417
DBI0.4720.6510.7320.7260.8530.8560.9461.0601.032
DI+(e-2)2.1692.9792.8372.9003.4893.3323.5182.9713.916
CHI+(e4)4.6905.3295.3285.7795.4855.3835.1584.9254.846
FSI(e11)−4.610−6.600−6.902−7.003−6.808−6.619−6.414−6.209−6.039
XBI0.0610.1390.1570.1490.2170.1930.2280.2750.239
KI(e3)1.1312.5522.8882.7503.9953.5604.2015.0644.400
TI(e3)1.1312.5552.8922.7554.0043.5694.2135.0804.416
SCI+2.2413.1093.1614.2014.0124.5864.5744.4694.605
CWBI(e-3)0.4400.4850.5970.6510.9000.9291.1181.3471.333
WSJI0.2440.6810.2070.2440.4460.4910.7061.0191.020
PBMFI+(e6)4.1316.2471.7320.3190.5280.3250.1840.0090.005
SVFI+2.2042.4592.8673.1243.3073.5363.6973.7643.885
WLI1.2220.1850.2410.2310.2420.2440.2540.2810.307
Note: * denotes the actual cluster number of the image.
Table 9. Variations of the 16 CVIs with cluster numbers ranging from 2 to 10 for the HYDICE image.
Table 9. Variations of the 16 CVIs with cluster numbers ranging from 2 to 10 for the HYDICE image.
CVIsCluster Number
234 *5678910
PC+0.7520.7290.6690.6210.5870.5540.5410.5110.502
PE0.5730.7170.9221.1061.2461.3821.4711.5981.656
MPC+0.5040.5940.5580.5260.5050.4790.4750.4500.447
DBI0.8880.6690.7470.8240.8280.8990.8270.8880.888
DI+(e-2)1.1941.0701.1651.2361.0811.1901.8971.0981.089
CHI+(e4)1.0571.4941.4731.6181.6061.5081.6341.5871.594
FSI(e12)0.004−1.258−1.511−1.592−1.627−1.603−1.608−1.579−1.567
XBI0.1960.1050.1490.1680.2310.2580.2230.3150.260
KI(e3)2.0251.0841.5451.7332.3932.6712.3063.2582.694
TI(e3)2.0261.0851.5471.7372.3982.6782.3133.2682.704
SCI+0.3911.8781.9061.9972.1511.7331.9111.6911.901
CWBI(e-4)2.5381.7612.0172.3953.0823.5083.5854.5984.398
WSJI0.4221.0820.2290.2990.4820.6210.6711.1111.034
PBMFI+(e6)27.97828.2427.3111.4902.2741.2280.2500.6550.118
SVFI+1.5602.4902.9283.2983.5663.6894.2024.3524.349
WLI0.3900.2630.2880.3230.3880.4060.4750.4740.387
Note: * denotes the actual cluster number of the image.
Table 10. Variations of the 16 CVIs with cluster numbers ranging from 2 to 10 for the ROSIS image.
Table 10. Variations of the 16 CVIs with cluster numbers ranging from 2 to 10 for the ROSIS image.
CVIsCluster Number
2345 *678910
PC+0.7030.6640.6150.5940.5480.5040.4770.4610.443
PE0.6610.8781.0761.2041.3811.5411.6721.7751.874
MPC+0.4060.4950.4860.4920.4570.4210.4030.3930.381
DBI1.3050.7960.8940.8761.0011.4641.4051.3711.349
DI+(e-2)0.5800.6560.7110.6780.6760.6060.6280.5620.562
CHI+(e4)0.8341.3231.2171.1091.0000.9800.9050.8440.777
FSI(e11)2.514−1.127−1.939−2.526−2.678−2.627−2.636−2.671−2.637
XBI0.4270.1580.2730.2130.4810.7150.6660.6230.591
KI(e4)0.7900.2930.5060.3940.8901.3241.2331.1541.094
TI(e4)0.7900.2930.5060.3940.8911.3251.2341.1551.095
SCI+−0.0110.3160.2870.3900.2060.3330.1790.1470.591
CWBI(e-3)0.7310.4240.4930.4510.7000.8720.9461.0091.082
WSJI0.5020.7430.2610.2240.4320.6600.7890.9131.058
PBMFI+(e6)2.8831.8090.0680.0140.0100.0150.0070.0030.002
SVFI+0.5691.7202.1872.9583.2903.1633.6584.0114.288
WLI0.8590.4530.6240.6650.8720.6780.8480.8570.848
Note: * denotes the actual cluster number of the image.
Table 11. Variations of the 16 CVIs with cluster numbers ranging from 2 to 10 for the AVIRIS image.
Table 11. Variations of the 16 CVIs with cluster numbers ranging from 2 to 10 for the AVIRIS image.
CVIsCluster Number
2345 *678910
PC+0.8430.8380.8560.7320.7630.7000.6810.6610.631
PE0.3950.4510.4510.7600.6910.8890.9451.0141.135
MPC+0.6860.7570.8080.6650.7150.6500.6360.6190.590
DBI0.6900.5570.3830.6890.6170.8020.8640.9490.962
DI+(e-2)0.6151.5241.5670.8711.1981.2361.5231.2971.312
CHI+(e4)2.1052.9915.6864.6816.4605.6426.7166.2305.556
FSI(e11)−1.098−4.465−6.317−5.999−6.803−6.631−6.331−6.151−6.004
XBI0.1170.1250.0650.7970.5620.9480.6960.6430.829
KI(e4)0.1950.2100.1091.3350.9431.5891.1681.0801.394
TI(e4)0.1950.2100.1091.3380.9461.5951.1731.0861.401
SCI+1.1131.9584.7153.8445.6844.8135.3515.8696.006
CWBI(e-3)0.9820.5450.4071.2571.3672.0821.9381.9972.376
WSJI0.3600.6650.0790.2760.3350.7420.6760.7211.016
PBMFI+(e5)21.75114.5640.8554.1390.9501.2791.9250.9101.002
SVFI+1.9422.6533.4143.5443.6204.1293.4223.2203.007
WLI0.2350.2130.1270.1730.1490.2000.1580.1520.135
Note: * denotes the actual cluster number of the image.
Table 12. The optimal cluster numbers indicated by the CVIs by FCM for each remote sensing image.
Table 12. The optimal cluster numbers indicated by the CVIs by FCM for each remote sensing image.
ImagesK#PCPEMPCDBIDICHIFSISCI
Multispectral image
QuickBird33 *3 *3 *3 *3 *888
Landsat TM4222274 *4 *4 *
Landsat ETM+4322254 *37
GaoFen-15225 *5 *25 *5 *5 *
FLC1522325 *5 *5 *6
Hyperspectral image
Hyperion42222105610
HYDICE422338566
ROSIS 5424448610
AVIRIS5223343610
ImagesCXBIKITICWBIWSJIPBMFISVFIWLI
Multispectral image
QuickBird33 *3 *3 *3 *3 *3 *103 *
Landsat TM42224 *4 *282
Landsat ETM+422224 *2102
GaoFen-155 *5 *5 *35 *275 *
FLC1533335 *2102
Hyperspectral image
Hyperion422224 *3103
HYDICE433334 *393
ROSIS 533335 *2103
AVIRIS544444274
Note: K# in the second column denotes the actual cluster numbers of the images, * denotes that the actual cluster number of the image was correctly identified by the index.
Table 13. The optimal cluster numbers indicated by the CVIs by K-means for each remote sensing image.
Table 13. The optimal cluster numbers indicated by the CVIs by K-means for each remote sensing image.
ImagesK#PCPEMPCDBIDICHIFSISCI
Multispectral image
QuickBird3223 *3 *3 *977
Landsat TM4222284 *4 *4 *
Landsat ETM+4222294 *4 *9
GaoFen-15225 *5 *25 *5 *5 *
Hyperspectral image
Hyperion422227557
HYDICE422337775
ROSIS 522334399
AVIRIS5223343610
ImagesCXBIKITICWBIWSJIPBMFISVFIWLI
Multispectral image
QuickBird33 *3 *3 *3 *3 *3 *102
Landsat TM422224 *2102
Landsat ETM+422234 *2102
GaoFen-155 *5 *5 *5 *5 *295 *
Hyperspectral image
Hyperion422224 *3102
HYDICE455534 *2103
ROSIS 566635 *2103
AVIRIS544444274
Note: * denotes that the actual cluster number of the image was correctly identified by the index; results on FLC1 image were not included in the table in consideration of the relative lower classification accuracy.

Share and Cite

MDPI and ACS Style

Li, H.; Zhang, S.; Ding, X.; Zhang, C.; Dale, P. Performance Evaluation of Cluster Validity Indices (CVIs) on Multi/Hyperspectral Remote Sensing Datasets. Remote Sens. 2016, 8, 295. https://doi.org/10.3390/rs8040295

AMA Style

Li H, Zhang S, Ding X, Zhang C, Dale P. Performance Evaluation of Cluster Validity Indices (CVIs) on Multi/Hyperspectral Remote Sensing Datasets. Remote Sensing. 2016; 8(4):295. https://doi.org/10.3390/rs8040295

Chicago/Turabian Style

Li, Huapeng, Shuqing Zhang, Xiaohui Ding, Ce Zhang, and Patricia Dale. 2016. "Performance Evaluation of Cluster Validity Indices (CVIs) on Multi/Hyperspectral Remote Sensing Datasets" Remote Sensing 8, no. 4: 295. https://doi.org/10.3390/rs8040295

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop