Unsupervised Wavelet-Feature Correlation Ratio Markov Clustering Algorithm for Remotely Sensed Images

: The spectrums of one type of object under different conditions have the same features (up, down, protruding, concave) at the same spectral positions, which can be used as primary parameters to evaluate the difference among remotely sensed pixels. The wavelet-feature correlation ratio Markov clustering algorithm (WFCRMCA) for remotely sensed data is proposed based on an accurate description of abrupt spectral features and an optimized Markov clustering in the wavelet feather space. The peak points can be captured and identified by applying a wavelet transform to spectral data. The correlation ratio between two samples is a statistical calculation of the matched peak point positions on the wavelet feature within an adjustable spectrum domain or a range of wavelet scales. The evenly sampled data can be used to create class centers, depending on the correlation ratio threshold at each Markov step, accelerating the clustering speed by avoiding the computation of Euclidean distance for traditional clustering algorithms, such as K-means and ISODATA. Markov clustering applies several strategies, such as a simulated annealing method and gradually shrinking the clustering size, to control the clustering convergence. It can quickly obtain the best class centers at each clustering temperature. The experimental results of the Airborne Visible/Infrared Imaging Spectrometer (AVIRIS) and Thermal Mapping (TM) data have verified its acceptable clustering accuracy and high convergence velocity.


Introduction
Identifying suspected targets from remotely sensed data is paramount in everyday life and research.Researchers have extensively investigated numerous clustering algorithms, including cutting-edge technologies, for remotely sensed images, such as Airborne Visible/Infrared Imaging Spectrometer (AVIRIS) and Thermal Mapping (TM) data.However, several limitations exist in these algorithms.One widely used clustering algorithm is K-means clustering, which, unfortunately, cannot automatically determine the number of classes [1,2].Moreover, it exhibits slow convergence due to its reliance on minimal spatial distance [3,4].
The approaches ISODATA [5,6] and ISMC [7,8] can determine the class number through self-iteration.Nevertheless, the challenge lies in determining their parameters, particularly in adjusting distance parameters with changing dimensions.On the other hand, orthogonal projection classification suffers from projection fluctuation issues under the restriction of the number of bands [9,10].Cui introduced a feature extraction method that computes vectorized pixel values from a localized window, enhancing Bag-of-Words (BoW) performance.However, this approach may lead to a reduction in classification accuracy [11,12].Peng et al. proposed a graph-based structural deep spectral-spatial clustering network to sufficiently explore the structure information among pixels.They designed a self-expression-embedded multi-graph auto-encoder to explore high-order structure associations among pixels, thereby capturing robust spectral-spatial features and global clustering structure [13].
Appl.Sci.2024, 14, 767 2 of 14 Furthermore, Firat et al. developed a hybrid 3D residual spatial-spectral convolution network to extract deep spatio-spectral features using 3D CNN and ResNet architecture [14].Acharyya combined wavelet theory and neuro-fuzzy techniques for segmentation purposes [15,16].However, their feature extraction approach solely considers the absolute values of wavelet coefficients, neglecting the specific spectral patterns, and the computational requirements take time and effort.
A wavelet-feature correlation ratio Markov clustering algorithm (WFCRMCA) is proposed to differentiate the pixels according to the spectrum similarity among pixels.Of course, the spectrums of one object under different conditions are different.Still, they have the same features (up, down, protruding, concave, see Figure 1) at the same spectral positions, which are the main parameters used to evaluate the difference among remotely sensed pixels [17,18].Therefore, these characteristic positions can denote class features.Fortunately, band-pass wavelet filters can decompose data at different scales to detect these characteristics.
Appl.Sci.2024, 14, x FOR PEER REVIEW 2 of 14 structure associations among pixels, thereby capturing robust spectral-spatial features and global clustering structure [13].Furthermore, Firat et al. developed a hybrid 3D residual spatial-spectral convolution network to extract deep spatio-spectral features using 3D CNN and ResNet architecture [14].Acharyya combined wavelet theory and neuro-fuzzy techniques for segmentation purposes [15,16].However, their feature extraction approach solely considers the absolute values of wavelet coefficients, neglecting the specific spectral patterns, and the computational requirements take time and effort.
A wavelet-feature correlation ratio Markov clustering algorithm (WFCRMCA) is proposed to differentiate the pixels according to the spectrum similarity among pixels.Of course, the spectrums of one object under different conditions are different.Still, they have the same features (up, down, protruding, concave, see Figure 1) at the same spectral positions, which are the main parameters used to evaluate the difference among remotely sensed pixels [17,18].Therefore, these characteristic positions can denote class features.Fortunately, band-pass wavelet filters can decompose data at different scales to detect these characteristics.WFCRMCA can statistically control clustering accuracy by adjusting parameters such as Tstart, Tend, and Tstep.A new concept, correlation ratio (CR), is proposed to reflect the similarity between two wavelet-transformed samples.With an accurate description of the abrupt spectral features, wavelet correlation ratios can differentiate pixels along spectral dimensions.Expanding spectral bands of multi-spectral images increases the number of characteristic points to enrich the features of classes.WFCRMCA forms the clustering space and initial class centers with evenly sampled pixels.Without the initial parameter problem of the K-means algorithm, WFCRMCA can quickly reach the best class centers at each clustering temperature and obtain optimal class centers on the whole scope at high speed by gradually decreasing the clustering scale and temperature.Several theorems are provided and proved to strengthen the WFCRMCA in the Section 2. In the Section 3, WFCRMCA receives favorable results for clustering Landsat TM images and AVIRIS hyperspectral images.

Methods
Although the spectral curves of the same objects under different conditions are somewhat different, they have the same feature points (upward, downward, maximum, and minimum) at the same spectral positions (Figure 1).The WFCRMCA could detect abrupt signals through band-pass wavelet transform, such as crossing zero and extreme points.WFCRMCA can statistically control clustering accuracy by adjusting parameters such as T start , T end , and T step .A new concept, correlation ratio (CR), is proposed to reflect the similarity between two wavelet-transformed samples.With an accurate description of the abrupt spectral features, wavelet correlation ratios can differentiate pixels along spectral dimensions.Expanding spectral bands of multi-spectral images increases the number of characteristic points to enrich the features of classes.WFCRMCA forms the clustering space and initial class centers with evenly sampled pixels.Without the initial parameter problem of the K-means algorithm, WFCRMCA can quickly reach the best class centers at each clustering temperature and obtain optimal class centers on the whole scope at high speed by gradually decreasing the clustering scale and temperature.Several theorems are provided and proved to strengthen the WFCRMCA in the Section 2. In the Section 3, WFCRMCA receives favorable results for clustering Landsat TM images and AVIRIS hyperspectral images.

Methods
Although the spectral curves of the same objects under different conditions are somewhat different, they have the same feature points (upward, downward, maximum, and minimum) at the same spectral positions (Figure 1).The WFCRMCA could detect abrupt signals through band-pass wavelet transform, such as crossing zero and extreme points.But crossing the zero point cannot be ensured to be a pulse signal, and perhaps is a smoothly changed signal, so the extreme points between adjacent zero points are much more critical.The signs in spectral vector format are classified according to the priority of importance from low to high: downward, upward, protruding, and concave (Figure 2).
Appl.Sci.2024, 14, x FOR PEER REVIEW 3 of 14 But crossing the zero point cannot be ensured to be a pulse signal, and perhaps is a smoothly changed signal, so the extreme points between adjacent zero points are much more critical.The signs in spectral vector format are classified according to the priority of importance from low to high: downward, upward, protruding, and concave (Figure 2).1), [19], 1st derivative Gauss function θ(t)).For some remotely sensed images affected by too many mixed pixels, the position of critical points will probably deviate or have many little fluctuations, so WFCRMCA could eliminate unimportant signals by setting a maximum threshold and only clustering the partial minutia at a high-level scale.
Wavelet feature clustering algorithms only analyze minutia data by detecting and determining the positions of abrupt signals.Using a fast binary Mallet wavelet algorithm [20] in Equation (2) to extract wavelet coefficients, WFCRMCA can mark the upward-maximal points (Figure 2a and Figure 2a') and downward-minimal points (Figure 2b and Figure 2b') along the spectrum.WFCRMCA will overlook the weak signals if Tpeak is large enough, leading to a failure in identifying some valuable signs among hidden objects.
The WFCRMCA uses rij (correlation ratio), which works like a distance but not Euclidean distance as clustering criteria, to evaluate the difference between two spectral vectors on partial minutia.Equation (3) uses Scale2-scale minutia of Scale-scale wavelet coefficients to cluster, ti,k is the kth feature of ith sampled vector, N(•) is the number of feature positions that match criteria.WFCRMCA could use binary values to mark whether the position is valuable enough to attend clustering.When 1), [19], 1st derivative Gauss function θ(t)).For some remotely sensed images affected by too many mixed pixels, the position of critical points will probably deviate or have many little fluctuations, so WFCRMCA could eliminate unimportant signals by setting a maximum threshold and only clustering the partial minutia at a high-level scale.
Wavelet feature clustering algorithms only analyze minutia data by detecting and determining the positions of abrupt signals.Using a fast binary Mallet wavelet algorithm [20] in Equation ( 2) to extract wavelet coefficients, WFCRMCA can mark the upward-maximal points (Figures 2a and 2a') and downward-minimal points (Figures 2b and 2b') along the spectrum.WFCRMCA will overlook the weak signals if T peak is large enough, leading to a failure in identifying some valuable signs among hidden objects.
The WFCRMCA uses r ij (correlation ratio), which works like a distance but not Euclidean distance as clustering criteria, to evaluate the difference between two spectral vectors on partial minutia.Equation ( 3) uses S cale2 -scale minutia of S cale -scale wavelet coefficients to cluster, t i,k is the kth feature of ith sampled vector, N(•) is the number of feature positions that match criteria.WFCRMCA could use binary values to mark whether the position is valuable enough to attend clustering.When S cale2 = S cale = ⌊log 2 b⌋, the bit number attending clustering comparison is b∑ Appl.Sci.2024, 14, 767 4 of 14

Expanding Bands Method for Multi-Spectral Images
As the band number of multi-spectral images (TM images have only seven bands) is not high enough for the wavelet transform to extract efficient feature points, WFCRMCA expurgates the bands with great noise and expands the rest with 2nd order and nonlinear correlated functions so that WFCRMCA can detect more wavelet features.The expanding multi-spectral bands' method [9] is listed as follows.

1.
Second-order correlated bands include the auto-correlated bands ( ) and the cross-correlated bands ( B i B j b i,j=1, i̸ =j ).

2.
Nonlinear correlated bands include the bands stretched out by the square root ) and those stretched out by the logarithmic function ({logB i } b i=1 ).The bands created by ( 1) and ( 2), together with the 1st order bands, which are original ({B i } b i=1 ), assemble new remotely sensed data with (b 2 + 7b)/2 bands.

Markov Chain Clustering in Wavelet Feature Space
The wavelet-feature Markov clustering algorithm, i.e., WFCRMCA, first denoises the original data to make the spectral features more accurate, then uses a band-pass wavelet filter to detect all dot vectors for sharp points, including upward-maximal and downward-minimal points.As a result, simulated annealing Markov chain decomposition in state space, formed by evenly spaced sampled data, could realize the best centers at each temperature and sub-finest centers on the whole scope.
According to the peculiarity of simulated annealing Markov clustering, each clustering center is one state, and the space is a definite Markov state chain.If two classes (or states) merge, according to CR, it has nothing to do with other states.For example, for Markov chain I = {1, 2, • • • , n} in definite state space {X(n)}, if any two states communicate, they must be in the same class.Thus, the whole state space (pixels) could be separated into a few isolated classes according to transferred communication.T, which is defined as a threshold value of CR r ij , is used as an annealing temperature to control the clustering process.
Proof of Theorem 1.According to Chapman-Kolmogorov equation: During the k+1th iteration, T k+1 = T k − T step , where T step is the depressed step of T at each iteration.If then i and j are merged together.If T step is small enough (i.e., the temperature is reduced slowly), and i, j, l are absorbed in k+1th iteration, p ij = p il = p jl ≈ T step so that it could be supposed as in Figure 3 that absorbed into one class.T is the CR threshold of the kth iteration, and c1,…, i, … j, …, and cn are the created clustering centers of the kth iteration.Thus,   .
As m states communicate with each other, other m − 1 states could be seen as one state j.Let pii = pjj = x, pij = 1 -x; thus So state i is a frequently returned state.As m states communicate, the merged m states are frequently returned.□ Theorem 3. The sufficient and necessary condition of closed set C is that, for arbitrary elements  ∈ ,  ∉ , there exists  0,  1 (referring [21,22]).[21,22]).
Therefore, every state is frequently returned in the wavelet feature space at each temperature, and the number of isolated closed sets equals the number of classes.Then, the whole wavelet Markov chain feature state space has a decomposable expression that consists of several closed sets without overlap.

Adjustment of Clustering Centers
When two classes are merged whose correlation ratio rij is bigger than T, the numbers of each feature (including crossing zero part) are separately added up at the corresponding position ([0, b − 1], the number of wavelet coefficients is approximately b).In addition, their sample numbers are also added up separately.As m states communicate with each other, other m − 1 states could be seen as one state j.Let p ii = p jj = x, p ij = 1 -x; thus So state i is a frequently returned state.As m states communicate, the merged m states are frequently returned.□ Theorem 3. The sufficient and necessary condition of closed set C is that, for arbitrary elements i ∈ C, j / ∈ C, there exists p [21,22]).
Theorem 4. Definite states of Markov chain in wavelet feature space can be uniquely decomposed withoutoverlap into a definite number of frequently returned states, including closed sets C 1 , . .., C m and single dot sets C m+1 , . .., C n , existing: When h ̸ = g,(h, gϵ [1, n]), any state in C g cannot communicate with any state in C h (referring [21,22]).
Therefore, every state is frequently returned in the wavelet feature space at each temperature, and the number of isolated closed sets equals the number of classes.Then, the whole wavelet Markov chain feature state space has a decomposable expression that consists of several closed sets without overlap.

Adjustment of Clustering Centers
When two classes are merged whose correlation ratio r ij is bigger than T, the numbers of each feature (including crossing zero part) are separately added up at the corresponding position ([0, b − 1], the number of wavelet coefficients is approximately b).In addition, their sample numbers are also added up separately.
Similar to the traditional clustering method, reasonable adjustment of clustering centers is based on the statistic of intra-class features.For each position, the feature that occurs most frequently is chosen as the common feature of the new class, and then b common features will be created.If several features come up at the same frequency, the feature with the highest priority (for example, downward-minimal or concave point >upward-maximal or protruding point) is chosen.Then, among all the class centers merged into one new class at this iteration, one pixel with the biggest CR with common features as a new class center is chosen.According to Equations ( 4) and ( 5), NB c i l,k is the statistic number of feature k on the lth position in class i, and t l is the feature of the lth position, which can be downward (0), upward (1), protruding (2), and concave (3).R(c 1 ,c 2 ) is the correlation ratio between vector c 1 and c 2 , Z c i is the set of class centers absorbed by class i, and M c i is the common features of class i.
During the clustering process, many pixels with high similarity are merged, causing the number of class centers that will attend the following iterative clustering comparison to decrease sharply.As only newly created centers follow next-cycle clustering, WFCRMCA has a high clustering speed.The computational complexity of WFCRMCA is O(n), which is impressive in handling large-scale datasets.

Wavelet-Feature Markov Clustering Algorithm
Based on the preceding theoretical analysis, the WFCRMCA uses a simulated annealing technique to gradually decrease CR threshold T through Markov chain decomposition in wavelet feature space at each temperature, obtaining the best clustering centers of the whole space.Supposed that c i is the class center of class i, S ci is the pixel set of class i, C is the set of all classes, Z ci is the set of class centers absorbed by class i at the current temperature, N c is the class number, N s is the number of sampled pixels (initial class centers are sampled pixels s i , i ∈ [0, N s − 1]), R(c 1 ,c 2 ) is the CR between c 1 and c 2 , N(Z c i ) is the number of class centers absorbed by class i, N(S c i ) is the pixel number in class i, T start is the initial value of CR T, and T end is the lowest CR threshold.The detailed process of WFCRMCA is provided in the flow chart in Figure 4.The simulated annealing Markov chain decomposition clustering in wavelet feature space is listed as follows: 1.
Input parameters: S tepx and S tepy are the sampling distances along horizontal or vertical directions; b, m, n are, separately, the band number, column number, and row number of original remotely sensed images; S cale is the wavelet transform scale; S cale2 is the number of minutia scale attending clustering (i.e., S cale −S cale2 ~Scale minutia sections)
According to S tepx × S tepy sampling distance, sample the pixels and create N s sampled pixels evenly.

5.
Apply simulated annealing Markov state decomposition clustering to S cale -S cale2 ~Scale scale minutia sections of sampled data.
(a) Set initial temperature T as T start , the clustering signal standard T signal (ratio of intraclass sampled pixel number over the number of total sampled pixels) is 1.0, and each pixel is one class center (beginning with N s class centers).In the end, according to step b-e, apply Markov chain decomposition in state space to the wavelet features of the sampled pixels by gradually depressing the signal size.
(b) Make judgments to all present class centers.If class i is a significant signal in which the number of pixels is more prominent than T signal N s , move to the next class.Otherwise, search forward one by one for another class j whose size is smaller than T signal N s , and make clustering judgments between class j and i. (c) According to Equation (8), if the CR between the centers of two classes (i and j) meets the condition P ij = r ij -T > 0, then class j is absorbed into class i. Continue this process (b) until the last class is detected.4) and ( 5

Results
The WFCRMCA uses Microsoft Visual C++ language (Microsoft, Redmond, DC, USA) and basic libraries for the code of the proposed algorithm.The TM and AVIRIS data analysis demonstrate the merits and defects of the wavelet feature clustering algorithms, say, WFCRMCA.Classified pixels are shown in white.

Results
The WFCRMCA uses Microsoft Visual C++ language (Microsoft, Redmond, DC, USA) and basic libraries for the code of the proposed algorithm.The TM and AVIRIS data analysis demonstrate the merits and defects of the wavelet feature clustering algorithms, say, WFCRMCA.Classified pixels are shown in white.
It is supposed that T start = 0.95 during the discussion of the parameters' influence on Mississippi's clustering results.If only the original six bands are processed by twoscale wavelet decomposition, only four classes are created because the features need to be stronger.As the first iteration absorbs too many classes, the intra-class adjustment costs most of the time.The experiment also shows that second-order correlation expanded bands It is supposed that Tstart = 0.95 during the discussion of the parameters' influence on Mississippi's clustering results.If only the original six bands are processed by two-scale wavelet decomposition, only four classes are created because the features need to be stronger.As the first iteration absorbs too many classes, the intra-class adjustment costs most of the time.The experiment also shows that second-order correlation expanded bands (7th-27th) provide more class information, but nonlinear correlation developed bands (28th-39th) make the clustering results stable.
The expanding spectrum method increases data processing complexity; however, if there are only several classes, the clustering speed is low because the big class has to spend more time calculating the center.Therefore, this method will maintain the clustering at a stable speed for multi-spectral data.Table 1 shows that this method could identify the potential specific classes, leading to higher clustering accuracy.Figure 6 is the clustering result of the parameters in Table 2.It can be seen that class 1 is plow land or meadow (Figure 6a), class 2 is beach (Figure 6b), class 3 is river channel (Figure 6c), class 6 is dyke (Figure 6f), and class 9 is the slope on the bank (Figure 6i).The clustering results maintain significant signals and efficiently embody the minor signs.If the data are divided into 18 classes by the K-means algorithm, one iteration, on average, uses 60 s, so this clustering method, according to the features on the spectral curves of remotely sensed objects, is more flexible on parameter choice and has a quicker clustering speed than the standard clustering algorithm (such as K-means).The expanding spectrum method increases data processing complexity; however, if there are only several classes, the clustering speed is low because the big class has to spend more time calculating the center.Therefore, this method will maintain the clustering at a stable speed for multi-spectral data.Table 1 shows that this method could identify the potential specific classes, leading to higher clustering accuracy.Figure 6 is the clustering result of the parameters in Table 2.It can be seen that class 1 is plow land or meadow (Figure 6a), class 2 is beach (Figure 6b), class 3 is river channel (Figure 6c), class 6 is dyke (Figure 6f), and class 9 is the slope on the bank (Figure 6i).The clustering results maintain significant signals and efficiently embody the minor signs.If the data are divided into 18 classes by the K-means algorithm, one iteration, on average, uses 60 s, so this clustering method, according to the features on the spectral curves of remotely sensed objects, is more flexible on parameter choice and has a quicker clustering speed than the standard clustering algorithm (such as K-means).
In Table 3, when T peak = 0, the number of classes is 133.If T peak is increased, the number of classes nonlinearly reduces, and clustering time depresses accordingly.When T peak > 7, the number of classes begins to fluctuate, so the WFCRMCA usually chooses T peak = 5.0, which could realize a fairly accurate classification.
In Table 3, when Tpeak = 0, the number of classes is 133.If Tpeak is increased, the number of classes nonlinearly reduces, and clustering time depresses accordingly.When Tpeak > 7, the number of classes begins to fluctuate, so the WFCRMCA usually chooses Tpeak = 5.0, which could realize a fairly accurate classification.
In Table 4, with more minutiae attending clustering, the number of clustering classes increases sharply: two of five scale components cluster eight categories; obviously, that does not separate the objects.However, four-scale components cause objects to disperse and expand the class number.Depressing Tend could effectively decrease the class number.
If high three-scale or five-scale wavelet decompositions are chosen to attend clustering and Tpeak = 5.0, 17 classes are created (the main clustering results are seen in Figure 7), the time is 21s, and the division result is favorable; here, class 1 is the basin (Figure 7a), class 4 is for the mountain peaks (Figure 7d), and class 5 is the water body of the Sook Lake (Figure 7e).In Table 4, with more minutiae attending clustering, the number of clustering classes increases sharply: two of five scale components cluster eight categories; obviously, that does not separate the objects.However, four-scale components cause objects to disperse and expand the class number.Depressing T end could effectively decrease the class number.
If high three-scale or five-scale wavelet decompositions are chosen to attend clustering and T peak = 5.0, 17 classes are created (the main clustering results are seen in Figure 7), the time is 21s, and the division result is favorable; here, class 1 is the basin (Figure 7a), class 4 is for the mountain peaks (Figure 7d), and class 5 is the water body of the Sook Lake (Figure 7e).

Discussion
For remotely sensed data with a high density of mixed pixels, choosing partial minutia wavelet features in high-level scales could reduce the clustering difficulty caused by a significant amount of minutia, and this also somewhat applies blur to achieve ideal clustering results.Multi-scale classification from fine to coarse could be realized by this method.Furthermore, as the matching speed of abrupt point positions is very high, clustering time increases a little with the increment of referenced minutia.For example, multispectral data typically set Scale = Scale2; hyper-spectral data could set Scale2 = Scale − 2.
WFCRMCA applies 1D wavelet transformation on satellite spectral data.Wavelet transform can represent a signal in both time and frequency domains simultaneously.It

Discussion
For remotely sensed data with a high density of mixed pixels, choosing partial minutia wavelet features in high-level scales could reduce the clustering difficulty caused by a significant amount of minutia, and this also somewhat applies blur to achieve ideal clustering results.Multi-scale classification from fine to coarse could be realized by this method.Furthermore, as the matching speed of abrupt point positions is very high, clustering time increases a little with the increment of referenced minutia.For example, multi-spectral data typically set S cale = S cale2 ; hyper-spectral data could set S cale2 = S cale − 2.
WFCRMCA applies 1D wavelet transformation on satellite spectral data.Wavelet transform can represent a signal in both time and frequency domains simultaneously.It decomposes a signal into a set of wavelets that are localized in both time and frequency, allowing me to analyze the signal's time-localized features.Wavelet transform excels at capturing localized features and adaptability to non-stationary signals.However, the Fourier transform represents a signal in the frequency domain.It decomposes a signal into a sum of sinusoidal components of different frequencies, providing information about the signal's frequency content.It does not capture information about when these frequencies occur.Fourier transform is excellent for spectral analysis.
Ridgelet and curvelet transformation are well-known methods for high-dimensional image analysis, but wavelet transformation is better for 1D spectral feature extraction.In the ridgelet transform, ridgelets are adapted to higher-dimensional singularities, or singularities on curves in dimension two, singularities on surfaces in dimension three, and singularities on (n − 1) dimensional hypersurfaces in dimension n [28].The curvelet transform uses ridgelet transform as a component step, and it is good at 2D image reconstruction [29].The proposed WFCRMCA uses wavelet transform to analyze the 1D spectral data instead of 2D images.
WFCRMCA accelerates clustering speed during the clustering process.The calculation of CR only needs simply matching corresponding characteristic points without the timeconsuming floating-point measure of Euclidean distance [1,2].A great many sampled pixels with high similarity are clustered together during the clustering process, depressing the number of class centers attending clustering comparison; moreover, clustering centers of newly created classes are re-determined according to common features.So, along with the process of this algorithm, the clustering speed continues to increase.
This WFCRMCA only makes statistics of the number of each wavelet feature on every info-position as the class feature and chooses the best pixel as the clustering center but does not directly use the CR matrix to investigate the dependency degree between sampled pixels, resolving spatial complexity problems.
Gradually depressing clustering size could let both small and large signals embody efficiently, and too many noise signals are merged so that the WFCRMCA could detect the spatial position of noise signals.
The WFCRMCA approach can be applied to any spectral data to differentiate targets.It has demonstrated favorable performance for satellite multi-spectral images and superspectral images.The spectral analysis method has potential applications in other spectrum data from photoacoustic imaging, OCT, etc. Spectrum analysis of the spectrum data from photoacoustic imaging and OCT has no successful cases.However, the proposed method provides a possibility of enhancing the classification accuracy through spectrum analysis.
Though the Markov clustering method is not parallelizable in choosing clustering centers, it can provide optimal clustering centers of the wavelet coefficients at a high convergent velocity.After clustering centers are determined, WFCRMCA can cluster the larger dataset in parallel.
In the WFCRMCA, even though most parameters are stable and can be used in most cases, several parameters, such as T peak and S cale2 , still need to be adjusted manually to increase clustering accuracy for specific applications.It is easy for a user to try different values to derive the best parameters for their application.Future work will continue to focus on optimizing the parameters.
The proposed WFCRMCA method analyzes the spectra of one type of object under different conditions with the same features (up, down, protruding, concave) so that it can differentiate the targets.This method has the potential to integrate with other equally essential features to increase the accuracy of classification.Incorporating multiple features in clustering provides more possibilities and challenging topics for future research.

Conclusions
The wavelet-feature correlation ratio is used to depict the distance between two pixels by analyzing the wavelet features of spectral curves for remotely sensed data.Based on the particularity of CR clustering, a wavelet-feature Markov clustering algorithm is proposed for searching the optimal class centers.After spatial data are evenly sampled, sharp points on the band-pass wavelet coefficients, including extreme points and crossing zero points, are captured and used for clustering matching.WFCRMCA accelerates clustering speed by avoiding the time-consuming Euclidean distance calculation used for general clustering algorithms.For multi-spectral data, nonlinear correlation expanded bands provide more class information than second-order correlation developed bands.Markov clustering based on simulated annealing realizes fast clustering convergence at each temperature.WFCRMCA can enhance classification accuracy through spectrum analysis for other applications with spectrum data.

Figure 1 .
Figure 1.Five points at different spatial positions within the same class have the same features at the exact spectral locations.

Figure 1 .
Figure 1.Five points at different spatial positions within the same class have the same features at the exact spectral locations.

Figure 2 .
Figure 2. The wavelet band-pass filter and four kinds of abrupt signals.(a-d) are the four critical signals: upward-maximal point, downward-minimal point, protruding crossing zero point, and concave crossing zero point.ψ(t) is the band-pass wavelet filter, (a'-d') are the output of the four signals through the wavelet filter.

Figure 2
Figure 2 is the result of four kinds of abrupt signals processed by ψ(t) (Equation (1),[19], 1st derivative Gauss function θ(t)).For some remotely sensed images affected by too many mixed pixels, the position of critical points will probably deviate or have many little fluctuations, so WFCRMCA could eliminate unimportant signals by setting a maximum threshold and only clustering the partial minutia at a high-level scale.

Figure 2 .
Figure 2. The wavelet band-pass filter and four kinds of abrupt signals.(a-d) are the four critical signals: upward-maximal point, downward-minimal point, protruding crossing zero point, and concave crossing zero point.ψ(t) is the band-pass wavelet filter, (a'-d') are the output of the four signals through the wavelet filter.

Figure 2
Figure 2 is the result of four kinds of abrupt signals processed by ψ(t) (Equation (1),[19], 1st derivative Gauss function θ(t)).For some remotely sensed images affected by too many mixed pixels, the position of critical points will probably deviate or have many little fluctuations, so WFCRMCA could eliminate unimportant signals by setting a maximum threshold and only clustering the partial minutia at a high-level scale.

Definition 1 .Theorem 1 .
If P 1 ij = r ij − T > 0 for states i and j, they have one-step transferred communication denoted as i 1 ↔ j.Communication can be transferred.If i m ↔ k and k n

Figure 3 .
Figure 3. (a) Closed set composed of five states.(b) Closed set with two states.

Theorem 4 .
Definite states of Markov chain in wavelet feature space can be uniquely decomposed without overlap into a definite number of frequently returned states, including closed sets C1, …, Cm and single dot sets Cm+1, …, Cn, existing: 1.Any two states in Ch ℎ 1,  are communicated.2. When ℎ , ℎ,  1,  , any state in Cg cannot communicate with any state in Ch (referring

Figure 3 .
Figure 3. (a) Closed set composed of five states.(b) Closed set with two states.
), re-adjust the newly created centers: among all the class centers merged into one new class at this iteration, choose one pixel with the biggest CR with common features as a new class center.(e) Let T = T − T step decrease clustering temperature, and T signal = T signal /2 reduce clustering size.Repeat steps (a)-(d) until T is reduced to the appointed small signal threshold T end or the set class number is reached.6.According to the clustering centers created by(5), each pixel is clustered into one class whose center has the maximal CR.

Figure 5 .
Figure 5. (a) Mississippi TM 4th band image after gray balance.(b) Sook Lake AVIRIS 60th band image after gray balance.

Figure 6 .
Figure 6.Mississippi TM image, WFCRMCA clustering results: (a-i) are the eight significant signals.

Figure 7 .
Figure 7. Sook Lake AVIRIS image WFCRMCA clustering result.(a-f) are the seven significant signals.

Figure 7 .
Figure 7. Sook Lake AVIRIS image WFCRMCA clustering result.(a-f) are the seven significant signals.
the k+1th iteration,    , where Tstep is the depressed step of T at , so  ↔ , then i and j are merged together.