1. Introduction
Hyperspectral imaging has attained immense popularity in remote sensing community in recent years owing to its high accuracy in classification and objbfect identification from remotely sensed images. Diverse application such as environmental studies [
1], agricultural studies [
2,
3], mineral mapping [
4], surveillance employ remotely sensed hyperspectral images. Hyperspectral images record the image intensity at several bands over the electromagnetic region [
5]. The inclusion of detailed spectral information about a considerable number of spectral bands increases the discriminative ability of the imaging technology leading to higher accuracy in target detection, classification and object identification [
6]. Hyperspectral imaging has been found to be very useful in identifying different objects from satellite-borne images. The object identification essentially employs the spectral unmixing method, which essentially estimates the reflectance profile of the spectrally distinct materials or endmembers. The reflectance pattern obtained from an image pixel is the resultant of reflectance profile of multiple signal sources or endmembers due to the poor spatial resolution of the imaging sensors. Spectral unmixing methods necessarily estimate the reflectance pattern of endmembers present in the image and compute its fractional abundance. Traditional unmixing methods involve three stages estimation of the number of endmembers, endmember estimation and calculation of abundance of endmembers [
7].
Predominantly hyperspectral unmixing can be broadly classified into two categories unsupervised and semi-supervised unmixing [
8] according to the availability of spectral library. Unsupervised unmixing methods identify the endmember and abundance matrix from the data itself. Whereas, semi-supervised approach considers the spectral library as the endmember matrix and computes the abundance matrix of the library endmembers. In recent years, semi-supervised unmixing strategy [
9,
10] has gained prevalence as application specific spectral libraries are available due to the rapid increase in MEMS-based optics. Dictionary pruning process identifies a smaller subset of the spectral library that can represent the image.
Traditional unsupervised unmixing methods predominantly employ convex geometric, non-negative matrix factorization or independent component analysis strategy to estimate the endmembers. The convex geometry based endmember estimation approaches include vertex component analysis [
11], pixel purity index [
12], convex cone analysis [
13], minimum volume enclosing simplex [
14,
15], minimum volume simplex analysis [
16], iterated constrained endmember extraction [
17], simplex growing algorithm (SGA) [
18]. Independent component analysis (ICA) based endmember estimation methods include [
19,
20,
21,
22]. Non-negative matrix factorization approaches for estimate the endmembers as well as abundance simultaneous incorporating regularization terms like low-rank constraints, total variation. The notable NMF based unmixing methods such as Huang et al. [
23], Wang et al. [
24], Tsinos et al. [
25], Arngren et al. [
26], Jia et al. [
27], Huck et al. [
28], Zhang et al. [
29] employ different regularization terms to constrain the solution. However, unsupervised unmixing methods can produce satisfactory performance only when some of the image pixels contain dominant endmembers.
Many semi-supervised unmixing methods aim at computing the sparse abundance matrix assuming the spectral library as the endmember matrix. Popular sparse unmixing methods like-variable splitting and augmented Lagrangian (SUnSAL) [
30] employed
sparsity term, collaborative SUnSAL algorithm [
31] combined collaborative sparse regression with the sparsity promoting term, whereas, SUnSAL-TV [
32] introduced a total variation regularization term in the sparse unmixing. Among the sparse unmixing methods for abundance estimation robust sparse unmixing [
33,
34] method incorporates a redundant regularization term to account for endmember variability, joint local abundance method [
35] performs local unmixing by exploiting structural information of image, co-evolutionary approach [
36] formulates a multi-objective strategy and minimize it by evolutionary algorithm. Other works such as Feng et al. [
37] proposed a spatial regularization framework which employs maximum a posteriori estimation, Themelis et al. [
38] introduced a hierarchical Bayesian model based sparse unmixing method, Zhang et al. [
39] transform data in framelet domain and maximize the sparsity of the obtained abundance matrix, Zhu et al. [
40] proposed a correntropy maximization approach for sparse unmixing. Some recent works such as Li et al. [
34], Feng et al. [
41], Mei et al. [
42] used spatial information alongside spectral properties of the data. Since sparse unmixing consider the whole spectral library as endmember. The prevalent sparse unmixing methods mentioned above generate an abundance matrix which has lower level of sparsity.
Some library aided unmixing methods employ a pre-processing stage, which prunes the spectral library used. Prevalent dictionary pruning based unmixing methods include orthogonal matching pursuit (OMP) [
43], OMP Star [
44], subspace matching pursuit (SMP) [
45], compressive sampling matching pursuit (CoSaMP) [
46], simultaneous orthogonal matching pursuit (SOMP) [
47], MUSIC-collaborative sparse regression (MUSIC-CSR) [
48], robust MUSIC-dictionary aided sparse regression (RMUSIC-DANSER) [
49], sparse unmixing using spectral apriori information (SUnSPI) [
50], centralized collaborative unmixing [
51], deblurring and sparse unmixing [
52] regularized simultaneous forward–backward greedy algorithm (RSFoBa) [
53], nuclear norm approach [
54]. include a pruning stage. Other works such as Li et al. [
55] proposes a collaborative sparse regression approach which considers the non-linearity as an outlier and employs an inexact augmented Lagrangian method to solve the optimization problem. MUSIC-CSR algorithm [
48] identifies the signal subspace and its dimension by HySIME [
56] in the preliminary stage. The algorithm projects each library element on the signal subspace and identifies the signal components from the resulting projection error. Robust MUSIC algorithm (RMUSIC) [
48] proposes an improved noise robust version of the inversion process, which also accounts for the variability in the reflectance profile and the discrepancy in the reflectance profile between spectral library elements and the actual image endmembers. Greedy algorithms like OMP [
43], OMP star [
44], SOMP [
47], SMP [
45], CoSaMP [
46] find the best matching projections of multidimensional data onto an over-complete dictionary. However the above mentioned dictionary pruning algorithms have some inherent shortcomings, which are listed below
Researchers have proposed several sparse inversion approaches [
31] to compute abundance of the endmembers. Among the seminal works sparse unmixing method through variable splitting and augmented Lagrangian (SUnSAL) [
30] employed
sparsity term, collaborative SUnSAL algorithm [
31] added a collaborative sparse regression with the sparsity promoting term, whereas, SUnSAL-TV [
32] introduced a total variation regularization term in the sparse unmixing. Among the sparse unmixing methods for abundance estimation robust sparse unmixing [
33] method incorporates a redundant regularization term to account for endmember variability, joint local abundance method [
35] performs local unmixing by exploiting structural information of image, co-evolutionary approach [
36] formulates a multi-objective strategy and minimize it by an evolutionary optimization. Other works such as Feng et al. [
37] proposed a spatial regularization framework which employs maximum a posteriori estimation, Themelis et al. [
38] introduced a hierarchical Bayesian model based sparse unmixing method, Zhang et al. [
39] transform data in framelet domain where abundance sparsity is maximized, Zhu et al. [
40] proposed a correntropy maximization approach for sparse unmixing. Some recent works such as Li et al. [
34], Feng et al. [
41], Mei et al. [
42] used spatial information alongside spectral properties of the data.
In this paper, we propose a novel dictionary pruning approach, where we identify the optimum image endmembers employing popular PCA based dimensionality reduction. In this work, we have employed recursive PCA formulation to minimize the computational time significantly due to the repetitive computation of eigenvalue. We also include a compressive sensing based framework to reduce the mutual coherence of spectral library. The experimental results shown in the paper demonstrate that our proposed dictionary pruning is a faster and straight-forward unmixing method, which can identify the exact endmember set.
Overall the paper is organized into the following sections
Section 2 presents the signal model for linear unmixing and describes the existing algorithms,
Section 3 illustrates the proposed mutual coherence reduction strategy and PCA based dictionary pruning method,
Section 4 presents the results obtained on simulated as well as real images, whereas,
Section 5 includes the conclusion and presents future scope of the proposed work.
3. Proposed Dictionary Pruning Method
In this paper, we introduce two novel dictionary pruning algorithms PCA reconstruction error difference (PRER) and PCA reconstruction error ratio (PRER). Our proposed unmixing framework comprises of four stages noise removal, estimation of the number of endmembers, dictionary pruning and abundance computation. We include an additional mutual coherence reduction stage before unmixing for improving its performance. We utilize multi-linear regression for denoising [
56], Harsanyi Ferrand Chang Virtual Dimensionality (HFC-VD) [
57] for estimation of the number of endmembers along with a novel method for mutual coherence reduction. The mutual coherence reduction task have not been explored in hyperspectral sparse unmixing.
3.1. Noise Removal by Multi Linear Regression
Since efficient noise removal is pertinent to spectral unmixing we employ multilinear regression [
58] framework for noise removal because of its improved performance in the hyperspectral setting [
56]. This method estimates the noise present in the data by using the correlation between the consecutive spectral bands. The method models the reflectance pattern of a spectral band as a linear regressive model of other spectral bands, motivated by the high correlation between the consecutive bands.
The reflectance value of all pixels in the
i-th band can be represented by
where,
represent reflectance profile of the
i-th band;
is the regression coefficient;
is the reflectance of all bands except the
i-th band; and
represents noise in the
i-th band. The regression coefficient is calculated by
The noise in the
i-th band can be estimated as
The noise free image at the
i-th band can be obtained by
The noise free image obtained by the process leads to improved unmixing performance.
3.2. Estimation of the Number of Endmembers
State of the art algorithms for estimation of the number of endmembers include-Harsanyi-Ferrand Chang virtual dimensionality (HFC-VD) [
57], hyperspectral subspace identification by minimum error (HySIME) [
56], eigen GAP index [
59], eigen thresholding [
60], low rank subspace estimation [
61], entropy estimation of eigenvalue [
62], maximal orthogonal complement algorithm (MOCA) [
63], high-order statistics (HOS)-HFC [
64] and hyperspectral subspace identification by SURE [
65], convex geometric approach GENE-CH and GENE-AH [
66], Hubness phenomenon [
67] etc. In this paper, we employ HFC-VD algorithm [
57] for estimation of the number of endmembers due to its accuracy in the hyperspectral setting.
3.3. Mutual Coherence Reduction
Mutual coherence of a spectral library indicates the maximum similarity between any pair of library elements. The high mutual coherence of spectral library creates complications in library aided unmixing as dictionary pruning algorithms identify consider the library elements with similar reflectance profile as distinct endmembers. Identification of duplicate endmembers reduces sparsity level of the obtained abundance matrix. The mutual coherence of a spectral library of size
is computed by
Ideally, the performance of unmixing should remain relatively unaffected by the high mutual coherence of spectral library. Although researchers have attempted to address the problem of high mutual coherence of spectral library in sparse inversion problems and compressive sensing, its effect on hyperspectral unmixing and mutual coherence reduction task has not bee carried out in hyperspectral unmixing.
In this paper, we also introduce a compressive sensing method to reduce the mutual coherence of the spectral library used. The high mutual coherence of spectral libraries creates a challenge in the library based unmixing of hyperspectral image. Mutual coherence measure identifies the maximum degree of similarity between any pair of spectral library elements. A spectral library with high mutual coherence leads to the identification of multiple spectral library elements as single endmember.
The problem of mutual coherence reduction of dictionary or library arises in sparse inversion problems in compressive sensing setting. Compressive sensing aims at obtaining the sparsest solution to the linear system
in terms of
norm. Here,
represents the measurement data,
is the over-complete dictionary and
indicates the sparse coefficient vector. According to compressive sensing formulation the problem is written as
The criteria for obtaining the sparsest solution of the problem [
68] are displayed below
Under this condition,
is the sparsest solution. The low mutual coherence of dictionary facilitates the sparsest solution whereas, high mutual coherence of dictionary creates problems in pruning. Welch et al. [
69] derived a theoretical bound on mutual coherence of dictionary
D of size
. According to the bound, the minimum possible mutual coherence of the library is given by
Since the dictionary employed in the process (D) is fixed, the aim of mutual coherence reduction method is to estimate an optimum projection matrix K which leads to lower values of mutual coherence ().
The mutual coherence reduction method uses a random projection matrix K as the initial transformation matrix and obtains the transformed dictionary. The transformed dictionary is normalized such that the rows have unit norm.
In the mutual coherence reduction method, we minimize an alternate measure of mutual coherence termed as t-averaged mutual coherence, since, computation of mutual coherence is an NP-hard problem. Hence, we propose an alternate mutual coherence measure called t-averaged mutual coherence as this is computationally more affordable. we exploit the fact that the diagonal entries of a Gram matrix contains 1, when the library elements are normalized. The t-averaged mutual coherence [
70] term is calculated according to
We aim to minimize the mutual coherence term while satisfying the properties of Gram matrix. The mutual coherence reduction task is carried out according to the following steps in the first stage, we initialize the transformation matrix
K ,normalize the rows of spectral library and compute t-averaged mutual coherence of
M according to (
17). In the succeeding stage, we compute the Gram matrix and shrink its elements according to
The shrinking or thresholding operation performed by the aforementioned process makes the matrix
G a full-rank matrix. Hence, we reduce the rank of the matrix
G into
R by applying singular value-shrinkage. Compute the square root of
G according to
where
. We minimize
while satisfying the constraint
, which indicates that
S should be a close approximate of the updated library
. We employ adaptive direction method of multipliers (ADMM) [
71] based optimization framework, which identifies the transformation matrix that minimizes the mutual coherence
. The optimization method uses an indirect formulation for mutual coherence reduction is as follows
The problem is expressed according to the Lagrangian function as
Here, the second term limits the power of the transformed library
M. ADMM formulation employs a new slack variable and assumes that
where,
. ADMM framework solves the sub-problem
ADMM solution updates
K,
Z and
U according to
The transformation matrix P obtained by the process minimizes mutual coherence of the library. The algorithmic steps are clearly described in details in Algorithm 1.
Algorithm 1: Reduction of Mutual Coherence Reduction of Spectral Library |
Input: Spectral library with high mutual coherence |
Output: Spectral library with relatively lower mutual coherence |
Initialization: Select a random initial projection matrix |
1: Compute the transformed library |
2: Compute the Gram matrix |
3: Set the threshold value t |
4: Compute t-averaged mutual coherence according to Equation (17)
|
5: while The optimization problem Equation (20) not converged do |
6: Normalize M to unit length
|
7: Shrink the elements of G according to
|
8: Obtain the square root of the Gram matrix M according to |
9: Apply SVD on M and reduce the rank of M to m |
10: end while |
3.4. Dictionary Pruning by Recursive PCA
Any hyperspectral data lives in a substantially lower dimensional subspace, since, the data arises from a latent linear mixing process. The dimension of the subspace is close to the number of signal sources or intrinsic dimensionality of the data. Accurate identification of the intrinsic dimensionality is pivotal in dictionary pruning.
We identify the lower dimensional data subspace using Principal component analysis (PCA). Different signal processing and machine learning application have employed Principal component analysis (PCA) as a tool for dimensionality reduction. However, researchers have rarely exploited explored the possibility of employing PCA for dictionary pruning. PCA identifies a low dimensional signal subspace of dimension
d from the original data space (of dimension
D). These
d principal components correspond to the maximum variance of the data. The first principal component represents the maximum variance, and each succeeding component corresponds to the next highest variance under the constraint that it is orthogonal to the preceding components. first-
d PC’s obtained are statistically uncorrelated and orthogonal to each other. Rank
d PCA minimizes the least square error such that the transformed data has low rank
dSince, PCA is a data driven transformation method, both the transformed data () and the reconstruction error () depends solely on the retained dimension (d). However, the optimum reconstruction error corresponds to the numerical rank of the data.
3.4.1. Proposed PCA Reconstruction Error Ratio Criteria (PRER)
According to Craig’s unmixing criteria [
72], a hyperspectral data consisting of
P endmembers lies in a
-dimensional subspace obtained by PCA transformation. Transformation of the data into
-dimension leads to optimum reconstruction error and reducing the data further do not reduce the reconstruction error significantly.
We propose a dictionary pruning idea based on PCA reconstruction error ratio. In this approach, we append each library element with the data, obtain an augmented data and transform it to
-dimension. The augmented data
comprise of either
P endmembers or
endmembers. We identify the number of endmembers present indirectly from PCA reconstruction error obtained from
. Let,
represent the reconstruction error obtained after transforming
into
dimension using PCA. Intrinsic dimensionality or numerical rank of the augmented data
relies on the properties of the library element added
. The numerical value of the reconstruction error obtained after transforming the augmented data also
depends on the properties of
. If,
is an image endmember
is expected to be large, whereas, if
is not an actual image endmember
is lower. We propose an index called PCA reconstruction error ratio (PRER), which is expressed as
This index PRER has considerably lower numerical value for actual image endmembers and has a higher value for the other library elements. Hence, we consider PRER as a parameter-free indirect measure to identify the image endmembers. We present the detailed implementation of PRER based pruning in Algorithm 2.
3.4.2. Proposed PCA Reconstruction Error Difference Criteria (PRED)
Whenever a particular spectral library endmember (
) is appended with the data
X, the augmented data (
) lies in either
P dimensional or
-dimensional linear subspace, depending on whether the library endmember is a part of image data or not. In the first case, when the spectral library element is also an image endmember, the intrinsic dimension of the subspace is
otherwise, the intrinsic dimensionality is
P. In the first situation the reconstruction error
is low, in the other scenario,
is much higher. The difference in reconstruction error between actual data and appended data
gives a quantitative measure which indicates whether the spectral library endmembers are also present in the image. We present the algorithmic steps for PRED based library pruning in Algorithm 3.
Algorithm 2: PCA Reconstruction Error Ratio Criteria (PRER) for Dictionary Pruning |
Input: Hyperspectral image data , Spectral library , Number of endmembers |
Output: Index of the pruned library , Pruned library |
Initialization:
|
1: Transform the data into -dimension by PCA. Record PCA reconstruction error .
|
2: for and do |
3: Append each library element with the data matrix according to |
4: Calculate the reconstruction error by transforming the appended data into -dimension by PCA. Obtained reconstruction error |
5: Calculate PCA reconstruction error ratio criteria |
6:end for |
7: Consider the -elements corresponding to the minimum reconstruction error ratio as endmembers. Index of pruned library .
|
8: Pruned library |
9: return Index of the pruned library elements , pruned library |
Algorithm 3: Dictionary Pruning by PCA Reconstruction Error Difference Criteria (PRED) for Dictionary Pruning |
Input: Hyperspectral image data , Spectral library , Number of endmembers P |
Output: Index of the pruned library , Pruned library |
Initialization:
|
Transform the data into -dimension by PCA. PCA reconstruction error |
2: for and do |
Append each library element with the data matrix |
4: Transform the data into -dimension by PCA and record the reconstruction error |
Calculate the difference in reconstruction error |
6: end for |
Consider the P-elements corresponding to the minimum reconstruction error difference as image endmember index .
|
8: Obtain pruned library by |
return Index of the pruned library elements and Pruned library |
3.5. Recursive Principal Component Analysis
Our proposed library pruning methods PCA reconstruction error ratio (PRER) criteria and PCA reconstruction error ratio difference (PRED) rely on repeated computation of eigenpairs of the covariance matrix of the data. We incorporate a faster formulation to estimate the covariance matrix after augmenting a spectral library element according to rank one modification. Let, the covariance matrix of the appended data
be denoted by
. This covariance matrix after appending a row can be computed from the covariance matrix of the original data using the formula
We perform standard eigen decomposition on this modified covariance matrix, which reduces the computational runtime.
3.6. Abundance Computation
We employ SUnSAL-TV [
32] algorithm for abundance computation. Since, the hyperspectral image of any natural ground scene is smooth in the spatial domain, the abundance of the endmembers obtained by the unmixing method should also inherit the smoothness. This method exploits total variation of abundance along with sparsity and reconstruction error constraints. The overall formulation of this method is
Here, the first term indicates reconstruction error, whereas, the second term computes
sparsity and the final term indicate total variation. The total variation term
essentially represents the difference in neighbourhood pixels.