Fast Semi-Supervised Unmixing of Hyperspectral Image by Mutual Coherence Reduction and Recursive PCA

: Dictionary pruning step is often employed prior to the sparse unmixing process to improve the performance of library aided unmixing. This paper presents a novel recursive PCA approach for dictionary pruning of linearly mixed hyperspectral data motivated by the low-rank structure of a linearly mixed hyperspectral image. Further, we propose a mutual coherence reduction method for pre-unmixing to enhance the performance of pruning. In the pruning step we, identify the actual image endmembers utilizing the low-rank constraint. We obtain an augmented version of the data by appending each image endmember and compute PCA reconstruction error, which is a convex surrogate of matrix rank. We identify the pruned library elements according to PCA reconstruction error ratio (PRER) and PCA reconstruction error difference (PRED) and employ a recursive formulation for repeated PCA computation. Our proposed formulation identiﬁes the exact endmember set at an affordable computational requirement. Extensive simulated and real image experiments exhibit the efﬁcacy of the proposed algorithm in terms of its accuracy, computational complexity and noise performance.


Introduction
Hyperspectral imaging has attained immense popularity in remote sensing community in recent years owing to its high accuracy in classification and objbfect identification from remotely sensed images. Diverse application such as environmental studies [1], agricultural studies [2,3], mineral mapping [4], surveillance employ remotely sensed hyperspectral images. Hyperspectral images record the image intensity at several bands over the electromagnetic region [5]. The inclusion of detailed spectral information about a considerable number of spectral bands increases the discriminative ability of the imaging technology leading to higher accuracy in target detection, classification and object identification [6]. Hyperspectral imaging has been found to be very useful in identifying different objects from satellite-borne images. The object identification essentially employs the spectral unmixing method, which essentially estimates the reflectance profile of the spectrally distinct materials or endmembers. The reflectance pattern obtained from an image pixel is the resultant of reflectance profile of multiple signal sources or endmembers due to the poor spatial resolution of the imaging sensors. Spectral unmixing methods necessarily estimate the reflectance pattern of endmembers present in the image and compute its fractional abundance. Traditional unmixing methods involve three stages estimation of the number of endmembers, endmember estimation and calculation of abundance of endmembers [7].
Predominantly hyperspectral unmixing can be broadly classified into two categories unsupervised and semi-supervised unmixing [8] according to the availability of spectral library. Unsupervised unmixing methods identify the endmember and abundance matrix from the data itself. Whereas, semi-supervised approach considers the spectral library as the endmember matrix and computes the abundance matrix of the library endmembers. In recent years, semi-supervised unmixing strategy [9,10] has gained prevalence as application specific spectral libraries are available due to the rapid increase in MEMS-based optics. Dictionary pruning process identifies a smaller subset of the spectral library that can represent the image.
Many semi-supervised unmixing methods aim at computing the sparse abundance matrix assuming the spectral library as the endmember matrix. Popular sparse unmixing methods like-variable splitting and augmented Lagrangian (SUnSAL) [30] employed l 1 sparsity term, collaborative SUnSAL algorithm [31] combined collaborative sparse regression with the sparsity promoting term, whereas, SUnSAL-TV [32] introduced a total variation regularization term in the sparse unmixing. Among the sparse unmixing methods for abundance estimation robust sparse unmixing [33,34] method incorporates a redundant regularization term to account for endmember variability, joint local abundance method [35] performs local unmixing by exploiting structural information of image, co-evolutionary approach [36] formulates a multi-objective strategy and minimize it by evolutionary algorithm. Other works such as Feng et al. [37] proposed a spatial regularization framework which employs maximum a posteriori estimation, Themelis et al. [38] introduced a hierarchical Bayesian model based sparse unmixing method, Zhang et al. [39] transform data in framelet domain and maximize the sparsity of the obtained abundance matrix, Zhu et al. [40] proposed a correntropy maximization approach for sparse unmixing. Some recent works such as Li et al. [34], Feng et al. [41], Mei et al. [42] used spatial information alongside spectral properties of the data. Since sparse unmixing consider the whole spectral library as endmember. The prevalent sparse unmixing methods mentioned above generate an abundance matrix which has lower level of sparsity.
Some library aided unmixing methods employ a pre-processing stage, which prunes the spectral library used. Prevalent dictionary pruning based unmixing methods include orthogonal matching pursuit (OMP) [43], OMP Star [44], subspace matching pursuit (SMP) [45], compressive sampling matching pursuit (CoSaMP) [46], simultaneous orthogonal matching pursuit (SOMP) [47], MUSIC-collaborative sparse regression (MUSIC-CSR) [48], robust MUSIC-dictionary aided sparse regression (RMUSIC-DANSER) [49], sparse unmixing using spectral apriori information (SUnSPI) [50], centralized collaborative unmixing [51], deblurring and sparse unmixing [52] regularized simultaneous forward-backward greedy algorithm (RSFoBa) [53], nuclear norm approach [54]. include a pruning stage. Other works such as Li et al. [55] proposes a collaborative sparse regression approach which considers the non-linearity as an outlier and employs an inexact augmented Lagrangian method to solve the optimization problem. MUSIC-CSR algorithm [48] identifies the signal subspace and its dimension by HySIME [56] in the preliminary stage. The algorithm projects each library element on the signal subspace and identifies the signal components from the resulting projection error. Robust MUSIC algorithm (RMUSIC) [48] proposes an improved noise robust version of the inversion process, which also accounts for the variability in the reflectance profile and the discrepancy in the reflectance profile between spectral library elements and the actual image endmembers. Greedy algorithms like OMP [43], OMP star [44], SOMP [47], SMP [45], CoSaMP [46] find the best matching projections of multidimensional data onto an over-complete dictionary. However the above mentioned dictionary pruning algorithms have some inherent shortcomings, which are listed below

•
The size of the pruned library for algorithms like OMP [43], SMP [45], RSFoBa [53], CoSaMP [46], SUnSPI [50] is much higher compared to the actual number of endmembers. • Some algorithms require high computational time. • These algorithms tend to perform poorly when the mutual coherence of library is high.
Researchers have proposed several sparse inversion approaches [31] to compute abundance of the endmembers. Among the seminal works sparse unmixing method through variable splitting and augmented Lagrangian (SUnSAL) [30] employed l 1 sparsity term, collaborative SUnSAL algorithm [31] added a collaborative sparse regression with the sparsity promoting term, whereas, SUnSAL-TV [32] introduced a total variation regularization term in the sparse unmixing. Among the sparse unmixing methods for abundance estimation robust sparse unmixing [33] method incorporates a redundant regularization term to account for endmember variability, joint local abundance method [35] performs local unmixing by exploiting structural information of image, co-evolutionary approach [36] formulates a multi-objective strategy and minimize it by an evolutionary optimization. Other works such as Feng et al. [37] proposed a spatial regularization framework which employs maximum a posteriori estimation, Themelis et al. [38] introduced a hierarchical Bayesian model based sparse unmixing method, Zhang et al. [39] transform data in framelet domain where abundance sparsity is maximized, Zhu et al. [40] proposed a correntropy maximization approach for sparse unmixing. Some recent works such as Li et al. [34], Feng et al. [41], Mei et al. [42] used spatial information alongside spectral properties of the data.
In this paper, we propose a novel dictionary pruning approach, where we identify the optimum image endmembers employing popular PCA based dimensionality reduction. In this work, we have employed recursive PCA formulation to minimize the computational time significantly due to the repetitive computation of eigenvalue. We also include a compressive sensing based framework to reduce the mutual coherence of spectral library. The experimental results shown in the paper demonstrate that our proposed dictionary pruning is a faster and straight-forward unmixing method, which can identify the exact endmember set.
Overall the paper is organized into the following sections Section 2 presents the signal model for linear unmixing and describes the existing algorithms, Section 3 illustrates the proposed mutual coherence reduction strategy and PCA based dictionary pruning method, Section 4 presents the results obtained on simulated as well as real images, whereas, Section 5 includes the conclusion and presents future scope of the proposed work.

Signal Model for Linear Unmixing
According linear mixing model the spectral reflectance profile of the i-th pixel is written as where, a i denotes abundance of all endmembers in the i-th pixel. S = [s 1 , s 2 , · · · , s P ] is the endmember matrix which contains spectral signature of P endmembers. w i represents the noise present in the i-th pixel. The whole image X = [x 1 , x 2 , · · · , x N ] consisting of N pixels is represented in matrix form as The abundance values satisfies abundance non-negative constraint (ANC) and abundance sum to one constraint (ASC). ANC enunciate that abundance values are non-negative, whereas, ASC indicate that abundance vectors of a pixel sum to one. These constraints are expressed as

Semi-Supervised Unmixing
Semi-supervised unmixing algorithms consider the whole spectral library as endmember matrix and aims to estimate abundance of the spectral library using sparse inversion. Since the spectral library employed is over-complete, the obtained abundance matrix has higher levels of sparsity, which makes abundance estimation a sparse inversion problem, which represents the data as a sparse linear mixture of the library according to where, the hyperspectral image X = [x 1 , x 2 , · · · , x N ] comprises of N pixels and, the spectral library comprises of reflectance pattern of K elements; M R N×K represents abundance matrix; and W R N×L is the noise and residual term; Sparse unmixing algorithms obtain an abundance matrix M which leads to minimum reconstruction error while maximizing sparsity and satisfying other constraints where, 0 ≤ q ≤ 1 Here, the first term represents reconstruction error, whereas, the second term indicates sparsity of the obtained abundance matrix.

Dictionary Pruning
A hyperspectral data is represented as a mixture of pruned library as The pruned libraryD = d 1 ,d 2 , · · · ,d R contains R elements; andM = [m 1 ,m 2 , · · · ,m R ] is the estimated abundance matrix; The pruned library comprises selected atoms from the spectral library (D ⊂ D) which can represent the image in a compact formulation. Ideally, size of the pruned library be closer to the actual number of endmembers (R ≈ P) and R = P means exact match, which us the aim of ideal dictionary pruning based semi-blind unmixing algorithms.
However, these dictionary pruning algorithms have some inherent shortcomings, which are listed below

•
The size of the pruned library for algorithms like OMP [43], OMP star [44], SMP [45], RSFoBa [53], CoSaMP [46], SUnSPI [50] is much higher compared to the actual number of endmembers. • Some algorithms require high computational time. • These algorithms tend to perform poorly when the mutual coherence of library is high.
The mutual coherence of spectral library is the maximum cosine angle distance between any two spectral library elements. The mutual coherence of spectral library is defined as The value lies in the range [0, 1] and higher mutual coherence indicates higher similarity between multiple atoms of the spectral library. High mutual coherence leads to the identification of endmembers with similar reflectance pattern as separately. Mutual coherence reduction leads to better dictionary pruning performance.

Proposed Dictionary Pruning Method
In this paper, we introduce two novel dictionary pruning algorithms PCA reconstruction error difference (PRER) and PCA reconstruction error ratio (PRER). Our proposed unmixing framework comprises of four stages noise removal, estimation of the number of endmembers, dictionary pruning and abundance computation. We include an additional mutual coherence reduction stage before unmixing for improving its performance. We utilize multi-linear regression for denoising [56], Harsanyi Ferrand Chang Virtual Dimensionality (HFC-VD) [57] for estimation of the number of endmembers along with a novel method for mutual coherence reduction. The mutual coherence reduction task have not been explored in hyperspectral sparse unmixing.

Noise Removal by Multi Linear Regression
Since efficient noise removal is pertinent to spectral unmixing we employ multilinear regression [58] framework for noise removal because of its improved performance in the hyperspectral setting [56]. This method estimates the noise present in the data by using the correlation between the consecutive spectral bands. The method models the reflectance pattern of a spectral band as a linear regressive model of other spectral bands, motivated by the high correlation between the consecutive bands.
The reflectance value of all pixels in the i-th band can be represented by where, x :,i represent reflectance profile of the i-th band; β i is the regression coefficient; Y σ i = [x :,1 , x :,2 , · · · , x :,i−1 , x :,i+1 , · · · , x :,L ] is the reflectance of all bands except the i-th band; and ξ :,i represents noise in the i-th band. The regression coefficient is calculated bỹ The noise in the i-th band can be estimated as The noise free image at the i-th band can be obtained bỹ The noise free image obtained by the process leads to improved unmixing performance.

Mutual Coherence Reduction
Mutual coherence of a spectral library indicates the maximum similarity between any pair of library elements. The high mutual coherence of spectral library creates complications in library aided unmixing as dictionary pruning algorithms identify consider the library elements with similar reflectance profile as distinct endmembers. Identification of duplicate endmembers reduces sparsity level of the obtained abundance matrix. The mutual coherence of a spectral library of size K × L is computed by Ideally, the performance of unmixing should remain relatively unaffected by the high mutual coherence of spectral library. Although researchers have attempted to address the problem of high mutual coherence of spectral library in sparse inversion problems and compressive sensing, its effect on hyperspectral unmixing and mutual coherence reduction task has not bee carried out in hyperspectral unmixing.
In this paper, we also introduce a compressive sensing method to reduce the mutual coherence of the spectral library used. The high mutual coherence of spectral libraries creates a challenge in the library based unmixing of hyperspectral image. Mutual coherence measure identifies the maximum degree of similarity between any pair of spectral library elements. A spectral library with high mutual coherence leads to the identification of multiple spectral library elements as single endmember.
The problem of mutual coherence reduction of dictionary or library arises in sparse inversion problems in compressive sensing setting. Compressive sensing aims at obtaining the sparsest solution to the linear system in terms of L 0 norm. Here, x ∈ R n represents the measurement data, D ∈ R n×p is the over-complete dictionary and α ∈ R p indicates the sparse coefficient vector. According to compressive sensing formulation the problem is written as The criteria for obtaining the sparsest solution of the problem [68] are displayed below Under this condition, α is the sparsest solution. The low mutual coherence of dictionary facilitates the sparsest solution whereas, high mutual coherence of dictionary creates problems in pruning.
Welch et al. [69] derived a theoretical bound on mutual coherence of dictionary D of size m × p. According to the bound, the minimum possible mutual coherence of the library is given by Since the dictionary employed in the process (D) is fixed, the aim of mutual coherence reduction method is to estimate an optimum projection matrix K which leads to lower values of mutual coherence (µ(M)).
The mutual coherence reduction method uses a random projection matrix K as the initial transformation matrix and obtains the transformed dictionary. The transformed dictionary M = KD is normalized such that the rows have unit norm.
In the mutual coherence reduction method, we minimize an alternate measure of mutual coherence termed as t-averaged mutual coherence, since, computation of mutual coherence is an NP-hard problem. Hence, we propose an alternate mutual coherence measure called t-averaged mutual coherence as this is computationally more affordable. we exploit the fact that the diagonal entries of a Gram matrix contains 1, when the library elements are normalized. The t-averaged mutual coherence [70] term is calculated according to We aim to minimize the mutual coherence term while satisfying the properties of Gram matrix. The mutual coherence reduction task is carried out according to the following steps in the first stage, we initialize the transformation matrix K ,normalize the rows of spectral library and compute t-averaged mutual coherence of M according to (17). In the succeeding stage, we compute the Gram matrix and shrink its elements according to The shrinking or thresholding operation performed by the aforementioned process makes the matrix G a full-rank matrix. Hence, we reduce the rank of the matrix G into R by applying singular value-shrinkage. Compute the square root of G according to where S ∈ R L×R . We minimize µ t (KD) while satisfying the constraint S − KD 2 2 ≤ ξ, which indicates that S should be a close approximate of the updated library KD. We employ adaptive direction method of multipliers (ADMM) [71] based optimization framework, which identifies the transformation matrix that minimizes the mutual coherence µ t (PD). The optimization method uses an indirect formulation for mutual coherence reduction is as follows The problem is expressed according to the Lagrangian function as Here, the second term limits the power of the transformed library M. ADMM formulation employs a new slack variable and assumes that where, Z = K. ADMM framework solves the sub-problem ADMM solution updates K, Z and U according to The transformation matrix P obtained by the process minimizes mutual coherence of the library. The algorithmic steps are clearly described in details in Algorithm 1.

Algorithm 1: Reduction of Mutual Coherence Reduction of Spectral Library
Input: Spectral library with high mutual coherence D ∈ R R×L Output: Spectral library with relatively lower mutual coherence φ Initialization: Select a random initial projection matrix K ∈ R R×R Normalize M to unit length 7: Shrink the elements of G according to Obtain the square root of the Gram matrix M according to S T S = M 9: Apply SVD on M and reduce the rank of M to m 10: end while

Dictionary Pruning by Recursive PCA
Any hyperspectral data lives in a substantially lower dimensional subspace, since, the data arises from a latent linear mixing process. The dimension of the subspace is close to the number of signal sources or intrinsic dimensionality of the data. Accurate identification of the intrinsic dimensionality is pivotal in dictionary pruning.
We identify the lower dimensional data subspace using Principal component analysis (PCA). Different signal processing and machine learning application have employed Principal component analysis (PCA) as a tool for dimensionality reduction. However, researchers have rarely exploited explored the possibility of employing PCA for dictionary pruning. PCA identifies a low dimensional signal subspace of dimension d from the original data space (of dimension D). These d principal components correspond to the maximum variance of the data. The first principal component represents the maximum variance, and each succeeding component corresponds to the next highest variance under the constraint that it is orthogonal to the preceding components. first-d PC's obtained are statistically uncorrelated and orthogonal to each other. Rank d PCA minimizes the least square error such that the transformed data has low rank d Since, PCA is a data driven transformation method, both the transformed data (X) and the reconstruction error (E (d) ) depends solely on the retained dimension (d). However, the optimum reconstruction error corresponds to the numerical rank of the data.

Proposed PCA Reconstruction Error Ratio Criteria (PRER)
According to Craig's unmixing criteria [72], a hyperspectral data consisting of P endmembers lies in a P − 1-dimensional subspace obtained by PCA transformation. Transformation of the data into P − 1-dimension leads to optimum reconstruction error and reducing the data further do not reduce the reconstruction error significantly.
We propose a dictionary pruning idea based on PCA reconstruction error ratio. In this approach, we append each library element with the data, obtain an augmented data and transform it to P − 1-dimension. The augmented data Y i = [X; d i ] comprise of either P endmembers or P + 1 endmembers. We identify the number of endmembers present indirectly from PCA reconstruction error obtained from Y i . Let, E i (P − 1) represent the reconstruction error obtained after transforming Y i into P − 1 dimension using PCA. Intrinsic dimensionality or numerical rank of the augmented data Y i relies on the properties of the library element added d i . The numerical value of the reconstruction error obtained after transforming the augmented data also E i (P − 1) depends on the properties of d i . If, d i is an image endmember E i (P − 1) is expected to be large, whereas, if d i is not an actual image endmember E i (P − 1) is lower. We propose an index called PCA reconstruction error ratio (PRER), which is expressed as This index PRER has considerably lower numerical value for actual image endmembers and has a higher value for the other library elements. Hence, we consider PRER as a parameter-free indirect measure to identify the image endmembers. We present the detailed implementation of PRER based pruning in Algorithm 2.

Proposed PCA Reconstruction Error Difference Criteria (PRED)
Whenever a particular spectral library endmember (d i ) is appended with the data X, the augmented data (Y i = [X; d i ]) lies in either P dimensional or P − 1-dimensional linear subspace, depending on whether the library endmember is a part of image data or not. In the first case, when the spectral library element is also an image endmember, the intrinsic dimension of the subspace is P − 1 otherwise, the intrinsic dimensionality is P. In the first situation the reconstruction error E i (P − 1) is low, in the other scenario, E i (P − 1) is much higher. The difference in reconstruction error between actual data and appended data gives a quantitative measure which indicates whether the spectral library endmembers are also present in the image. We present the algorithmic steps for PRED based library pruning in Algorithm 3. Transform the data Y i into P − 1-dimension by PCA and record the reconstruction error Consider the P-elements corresponding to the minimum reconstruction error difference RE di f (i) as image endmember index φ.
8: Obtain pruned library byD = D φ return Index of the pruned library elements φ and Pruned libraryD

Recursive Principal Component Analysis
Our proposed library pruning methods PCA reconstruction error ratio (PRER) criteria and PCA reconstruction error ratio difference (PRED) rely on repeated computation of eigenpairs of the covariance matrix of the data. We incorporate a faster formulation to estimate the covariance matrix after augmenting a spectral library element according to rank one modification. Let, the covariance matrix of the appended data Y i be denoted byĈ i . This covariance matrix after appending a row can be computed from the covariance matrix of the original data using the formulâ We perform standard eigen decomposition on this modified covariance matrix, which reduces the computational runtime.

Abundance Computation
We employ SUnSAL-TV [32] algorithm for abundance computation. Since, the hyperspectral image of any natural ground scene is smooth in the spatial domain, the abundance of the endmembers obtained by the unmixing method should also inherit the smoothness. This method exploits total variation of abundance along with sparsity and reconstruction error constraints. The overall formulation of this method is Here, the first term indicates reconstruction error, whereas, the second term computes l 1 sparsity and the final term indicate total variation. The total variation term essentially represents the difference in neighbourhood pixels.

Results
We apply our proposed unmixing methods on a large number of synthetic and real images. We vary noise level, pixel purity level, the mutual coherence of the spectral library and number of endmembers in these synthetic image experiments.

Performance Measures
We evaluate the performance of the unmixing methods on two parameters signal to reconstruction error (SRE) and the probability of detection (Pr Det).

Signal to Reconstruction Error (SRE)
Signal to reconstruction error (SRE) denotes the relative power of reconstructed data with respect to the actual data. SRE = 10 log 10 X 2 X −X 2 (30) where,X is the hyperspectral data reconstructed by the unmixing or dictionary pruning algorithm.
Better unmixing leads to lower reconstruction error, which in turn increases SRE. •

Probability of Detection
The probability of detection defines the number of spectral library endmembers accurately selected according to the formula where Λ is the indices corresponding to the actual spectral library elements present in the image andΛ is the indices corresponding to the estimated spectral library elements of the pruned library.
Value of probability of detection lies in the range 0 ≤ Pr Det ≤ 1. The higher value specifies close match between actual and pruned library elements and exact match is represented by Pr Det = 1.

Synthetic Image Experiments
The synthetic images used in the experiments contain random endmembers from USGS spectral library and the abundance matrix was created according to Dirichlet distribution which satisfies ASC and ANC constraint. In the data A1 the number of endmembers is varied as five and ten and additive white Gaussian noise is added to the data. We alter the number of image pixels and maximum abundance of any endmember in a pixel in the synthetic data A2. In the data A3 we alter mutual coherence of the library and noise level simultaneously. In all these experiments, we perform mutual coherence reduction prior to unmixing.

Real Image Experiments
We used HYDICE Washington dc mall image (https://engineering.purdue.edu/biehl/MultiSpec/ hyperspectral.html) and HYDICE urban image (http://lesun.weebly.com/hyperspectral-data-set. html) to validate our proposed algorithms see the Figure 1. DC Mall hyperspectral image consists of 210 spectral bands which cover the electromagnetic range 400-2400 nm. We use a 188 spectral band version, which excludes the noisy and absorption bands present in the image. The ground truth results intimate that the image endmembers are covered in USGS spectral library. HyDICE urban image (http://lesun.weebly.com/hyperspectral-data-set.html) acquired by HYDICE sensor covers the electromagnetic spectrum range 400-2500 nm and comprise of 221 spectral bands. The image has a spatial size of 200 × 200 and consists of four endmembers as per ground-truth study [73]. However, the image contains noisy bands 1-4, 76, 87, 101-111, 136-153 and 198-210. We remove these bands before processing and use a 162 band version of the data for unmixing.
We plot the PCA reconstruction error ratio and PCA reconstruction error difference for each spectral library elements in Figure 2, which highlights that actual library endmembers lead to lower reconstruction error ratio and reconstruction error difference. Since the PRER and PRED values for the library elements are significantly lower compared to the other library elements, it is simple to identify the endmembers from these two parameters. We display the SRE comparison of data A1 and A2 on Figure 3, which highlights that PRER and PRED obtain significantly higher SRE compared to most of the methods. Figure 3a illustrates that PRER and PRED obtain relatively higher SRE for images with high levels of mixing (lower values of maximum abundance per pixel). Figure 3b suggests that our proposed PRER and PRED outperform the prevalent methods in presence of noise. However, we do not obtain satisfactory performance on extremely high noise levels. SRE values for most of the methods predictably decrease as noise level escalates. We show the abundance images corresponding to ground truth, PRER and PRED in Figures 4-6 respectively. The abundance images obtained by PRER and PRED are similar to the actual ground truth abundance image. This proves the potency of our proposed unmixing. We tabulate the probability of detection on data A1 and data A2 on Tables 1 and 2 respectively. The result displayed in Table 1 illustrates that our proposed algorithms obtain probability of detection equals to almost unity. The other result displayed in Table 2 shows that PRER and PRED result into a higher probability of detection in most of the situations. However, high SNR levels-20, 10 and 0 dB make it difficult to identify the exact set of image endmembers. Table 3 shows that PRER and PRED obtains almost unity probability of detection even for a dictionary with high mutual coherence. We present the probability of detection result for varied mutual coherence in Table 3. This result emphasizes that PRER and PRED obtains superior unmixing performance even in the presence of spectral library with high mutual coherence level.    Our proposed PRER and PRED based unmixing relies on basic operations like covariance matrix computation and eigen decomposition, it result into low computational complexity. We have employed Cupens divide and conqer algorithm [74] for eigen decomposition. This algorithm has a computational complexity of O n 2.3 . The formulation for performing rank 1 update has computational complexity of O n 3 . The overall complexity of the framework is O n 3 . The runtime comparison of PCA was reported in [75], whereas the computational complexity of robust PCA was reported in [76]. We compare the runtime performance on an i5 Core 2 Duo system having 8GB RAM. The runtime plot displayed in Figure 7 illustrates that SMP [45] is the fastest, closely followed by PRER or PRED. Although SMP [45] is computationally more efficient, its moderate noise performance and lower probability of detection make it unsuitable.

Conclusions
This paper introduces PCA as an alternative dictionary pruning method, which accurately estimates the exact spectral library endmember set if the noise level is under certain limit and the number of endmembers present in the image is accurately estimated. We incorporate a method to reduce the mutual coherence of spectral library, which improves the unmixing performance. We also present a recursive formulation for estimation of covariance matrix after rank one modification, which significantly improves the runtime performance of the proposed method.
Author Contributions: S.D. conceptualized and implemented the idea of recursive PCA approach for spectral library endmember selection. He prepared the manuscript. A.R. suggested the use of faster recursive formulation for computing covariance matrix of the augmented data matrix. He also contributed in giving the manuscript a compact format. A.K.D. helped improve the technical quality of the manuscript.