# HALS-based NMF with flexible constraints for hyperspectral unmixing

- Weishi Chen
^{1, 2}Email author and - Mireille Guillaume
^{1}

**2012**:54

https://doi.org/10.1186/1687-6180-2012-54

© Chen and Guillaume; licensee Springer. 2012

**Received: **29 June 2011

**Accepted: **5 March 2012

**Published: **5 March 2012

## Abstract

In this article, the hyperspectral unmixing problem is solved with the nonnegative matrix factorization (NMF) algorithm. The regularized criterion is minimized with a hierarchical alternating least squares (HALS) scheme. Under the HALS framework, four constraints are introduced to improve the unmixing accuracy, including the sum-to-unity constraint, the constraints for minimum spectral dispersion and maximum spatial dispersion, and the minimum volume constraint. The derived algorithm is called F-NMF, for NMF with flexible constraints. We experimentally compare F-NMF with different constraints and combined ones. We test the sensitivity and robustness of F-NMF to many parameters such as the purity level of endmembers, the number of endmembers and pixels, the SNR, the sparsity level of abundances, and the overestimation of endmembers. The proposed algorithm improves the results estimated by vertex component analysis. A comparative analysis on real data is included. The unmixing results given by a geometrical method, the simplex identification via split augmented Lagrangian and the F-NMF algorithms with combined constraints are compared, which shows the relative stability of F-NMF.

## Keywords

## 1. Introduction

Airborne hyperspectral sensors collect images in hundreds of narrow and contiguous spectral bands. Due to the limited spatial resolution of hyperspectral image (HSI), each observed pixel generally contains more than one material spectral signature. Hence, the hyperspectral unmixing, which decomposes a mixed pixel into a combination of pure material spectra known as endmembers, weighted by their corresponding abundance coefficients, is a challenging task.

**R**(

*L*×

*I*) be the matrix unfolded HSI, whose

*I*columns are the spectral pixels and the

*L*rows are the vectorial spectral band images. As

*N*is the related noise matrix, the linear spectral mixing model (LSMM) can be written as

The rows of **S** (*J* × *I*) are the abundance maps corresponding to the respective endmembers, whose spectra are located in the columns of **A** (*L* × *J*). *J* denotes the number of endmembers.

Basically, hyperspectral unmixing is a problem of blind source separation (BSS). However, compared with most BSS applications, the endmembers of HSI data are dependent and the elements in **A** and **S** are nonnegative, so the hyperspectral unmixing is beyond the reach of many BSS algorithms (e.g., independent component analysis-ICA) [1]. To fulfill these constraints, numerous special algorithms have been proposed to solve the hyperspectral unmixing problem under the LSMM assumption, including the approaches of convex geometry, Bayesian source separation, and nonnegative matrix factorization (NMF). The geometrical approaches first determine the endmembers and estimate the abundances in a second step, while the BSS and NMF-based approaches find the endmembers and the abundances simultaneously.

Geometrical approaches try to determine the vertices of the J-simplex enclosing the observed pixels, such as pixel purity index (PPI) [2], N-FINDR [3], vertex component analysis (VCA) [4]. The PPI algorithm projects every spectral vector onto skewers (large number of random vectors). The points corresponding to extremes, for each skewer direction, are stored and cumulated. The pixels with the highest scores are the purest ones. N-FINDR finds the set of pixels defining the largest volume within the data. VCA iteratively projects data onto a direction orthogonal and the endmembers correspond to the extreme of the projections. The issue of these approaches is to find extreme points within the data with the assumption of pure pixel of each endmember, which is always unsatisfactory for real hyperspectral data. Recently, the state-of-art reference algorithms MVSA [5], MVES [6], and the simplex identification via split augmented Lagrangian (SISAL) [7] have proposed various ways to find a minimum volume simplex, showing very good performances in the estimation of endmembers. Particularly, SISAL is able to unmix HSI data in the case of no pure pixel.

The geometrical approaches do not work well when the observed data are highly mixed, because there are not enough vectors in simplex facets. In these cases, the separation problem can be addressed in a Bayesian framework. Several Bayesian Positive Source Separation (BPSS) algorithms under positivity and sum-to-one constraints have recently been developed [8–10]. In [10], a discussion on the effectiveness of the sum-to-one constraint is given, showing that full constrained BPSS2 gives better results than BPSS for simulated data, while it is the contrary for the real OMEGA data, "due to nonlinearity in the radiative transfer and noise in the dataset in contradiction with the full additivity constraint". We think that it would be not the same with the proposed NMF-based algorithm, firstly because the full additivity is not a hard but a soft constraint, and second because the residual error RQE is able to represent measurement noise or model noise, and then the algorithm is quite robust for real data, which can contain nonlinear mixed terms. This can be seen by comparing the results of a geometrical algorithm like SISAL, very performant on simulated data, with the results obtained on Cuprite real data, which drops dramatically, while the NMF-based algorithms keep performing.

In the last decade, NMF has been a popular algorithm since Lee and Seung [11] investigated the properties of the algorithm and published some simple and useful algorithms for two types of factorizations. The NMF algorithm has broadly been used in text mining, image analysis, speech processing, and automatic control. The basic NMF problem consists of finding two nonnegative data matrices whose product approximates the mixed data in a chosen measure sense (e.g., the reconstruction quadratic error-RQE). However, the solution to NMF is not unique so various regularizations with prior knowledge should be taken into account to reduce the number of solutions. The sum-to-unity (STU) constraint is proposed in [12], which regularizes the RQE with a function of **S** to normalize the columns of it. The authors of [13] propose constraints based on two inherent characteristics of hyperspectral data: the spectral piecewise smoothness and spatial sparseness. In [14], a minimum volume constrained NMF (MVC-NMF) based on projected gradient (PG) optimization method is proposed, whose regularization term minimizes the simplex volume spanned by the endmembers. Other authors [15] propose a minimum distance constrained NMF (MDC-NMF), which consider the endmember distance instead of the volume of the estimated simplex. MDC-NMF makes a slight modification of the optimized algorithm used for MVC-NMF. MiniDisCo algorithm makes the assumption of minimum spectral dispersion for NMF regularization [16], and MDMD-NMF regularizes with minimum spectral dispersion and maximum spatial dispersion [17]. A new step-size estimation technique is proposed for the two algorithms to hasten the PG convergence.

The optimization algorithms and constraints on **A** and **S** are two main techniques for NMF-based hyperspectral unmixing. The authors of [18] propose a flexible hierarchical alternating least squares (HALS) algorithm with a set of local cost functions called alpha and beta divergences. The word "flexible" means the variation of the optimization algorithm. In this article, we propose an improved NMF algorithm with four constraints due to the characteristics of HSI, called the flexible NMF (F-NMF). The word "flexible" means the variation of constraints on **A** and **S**. F-NMF also uses the HALS update rules, significantly outperfoming the PG update rules in convergence speed. Actually, the novelty is both the combination of the constraints and the development of these constraints under HALS-based algorithm.

The rest of the article is organized as follows: Section 2 presents the basic NMF algorithm and the HALS update rules. In Section 3, we introduce four constraint functions and integrate them into the F-NMF algorithm. In Section 4, the comparison and analysis of the F-NMF with different constraints are given by processing various simulated HSIs. The algorithms are applied to real data in Section 5. The F-NMF algorithms are compared with SISAL, for the two algorithms are both able to unmix hyperspectral data in which the pure pixel assumption is violated. Finally, some conclusion closes the article.

## 2. NMF for hyperspectral unmixing

In this section, we first present the NMF problem and then the optimization algorithm used to solve it in this article.

### 2.1. NMF problem

where ||·||_{
F
} is the Frobenius (e.g., quadratic) norm.

### 2.2. HALS algorithms

In [19], the authors show that the HALS scheme works remarkably well in practice, outperforming, in most cases, the other optimization algorithms for NMF. In particular, it is proved to be locally more efficient [20] and shown to converge to a stationary point under some mild assumptions [21]. For these reasons, we choose HALS as the optimization technique.

for *k* = 1, 2,..., *J*. **A**_{
k
} (*L* × 1) is one endmember spectrum and **S**_{
k
} (1 × *I*) corresponds to its abundance fraction.

*k*= 1, 2,...,

*J*, where [δ][

_{0,1]}is to enforce every element δ

_{ ij }lies in [0,1], so

**A (A**

_{ k }

**)**does not depend on the other entries of the same column. By symmetry, the same property holds for each row of

**S**(

**S**

_{ k }). Thus, the detailed HALS algorithm is summarized as follows:

- (1)
Initialize

**A**and**S**with the VCA algorithm; - (2)
**for***i*= 1, 2,..., do

**for** k = 1, 2,..., *J*

Update **A**_{
k
} and **S**_{
k
} with the HALS update rules;

**end**

**until** the stop criteria is reached

*f*with all constraints will be proposed in Equation (18). The maximum number of iterations is always set high (e.g., 2000) to obtain accurate estimations. However, the overestimation of the iteration number induces time waste. Indeed, the RQE value slightly increases from certain iteration whereas the regularized

*f*keeps decreasing. Thus, the algorithm is stopped at this iteration when the RQE value goes to a minimum although the highest iteration number is not reached. The stop criteria is expressed as

## 3. NMF with flexible constraints

**A**and

**S**are both nonnegatives. Since the NMF solution is not unique, some prior knowledge on HSIs can be introduced to regularize the problem. A generic expression for the optimized function is

*D*(

**A**,

**S**) measures the difference between

**X**and

**AS**with respect to some norms. By substituting Equation (4) into (8) and using the RQE norm, the new optimized function

*f*is

In this section, we add four constraints for **A** and **S** to the function to improve the unmixing result. With all these constraints, the algorithm is called flexible NMF (F-NMF), based on HALS update rules.

### 3.1. STU constraint

**S**equal to 1. The STU constraint is defined as follows:

**1**

_{1I}is an (1 ×

*I*) vector of ones. The gradient derivation of

*D*

_{1}with respect to

**S**

_{ k }is

### 3.2. Maximum spatial dispersion constraint

*J*, we defined the maximum spatial dispersion constraint as follows:

*D*

_{2}with respect to

**S**

_{ k }is

### 3.3. Minimum spectral dispersion constraint

**A**, encouraging the variance of each endmember spectrum to be as low as possible. This dispersion constraint is to improve the shape estimation of flat endmember spectra. Consequently, if the estimation of some spectra is improved, the estimation of the other spectra involved in the mixture will also indirectly be improved due to the parameter interdependences. We define the minimum spectral dispersion constraint as

*D*

_{1}with respect to

**A**

_{ k }is

### 3.4. Minimum distance constraint

**A**is calculated as the constraint, which suffers from numerical instabilities [11]. Here, we choose the minimum distance constraint as a substitute in order to shrink the volume of the data enclosing the simplex. The distance is measured and summed up from every endmember to their centroid. This constraint is defined as

*D*

_{2}with respect to

**A**

_{ k }is

*f*with all these considerations are derived from (6), (9), (11), (13), (15), and (17). Thus

## 4. Simulations on synthetic data

In this section, we present a batch of simulations to quantitatively compare the F-NMF algorithms with different constraints. First, we present the used evaluation metrics. Then, we present the way we build simulated data. Finally, the experimental results of five F-NMF algorithms are given.

### 4.1. Evaluation metrics

- (1)To evaluate the abundance estimation, we define the abundance mean squared error (AME) as$\mathsf{\text{AME}}\left(\widehat{\mathbf{S}},\mathbf{S}\right)=\frac{1}{JI}{\u2225\widehat{\mathbf{S}}-\mathbf{S}\u2225}_{F}^{2}$(19)
- (2)To evaluate the endmember spectra estimation, we define the spectral mean squared error (SME) as$\mathsf{\text{SME}}\left(\widehat{\mathbf{A}},\mathbf{A}\right)=\frac{1}{LJ}{\u2225\widehat{\mathbf{A}}-\mathbf{A}\u2225}_{F}^{2}$(20)
- (3)To consider the global shape of the spectra, the spectral angle distance (SAD) is defined as$\mathsf{\text{SAD}}\left(\widehat{\mathbf{a}},\mathbf{a}\right)={cos}^{-1}\left(\frac{{\mathbf{a}}^{T}\widehat{\mathbf{a}}}{\sqrt{{\widehat{\mathbf{a}}}^{T}\widehat{\mathbf{a}}}\sqrt{{\mathbf{a}}^{T}\mathbf{a}}}\right)$(21)

where **a** is the true spectral vector and $\widehat{\mathbf{a}}$is its estimate.

### 4.2. Synthetic data

The HSI synthesis process is in three steps corresponding to the matrices **A**, **S**, and the noise matrix **N**.

First, the *J* endmember spectra are randomly selected among the U.S. Geological Survey (USGS) spectral library. The selected 224-channel spectra constitute the columns of the matrix **A**.

Then, the *J*-element column vector in **S** is generated following a Dirichlet pdf, with parameters equal to 1. The element maximal value of each column is controlled by a threshold *ξ* (0 <*ξ* ≤ 1). This operation allows one to control the mixing or purity level of the data. In particular, the image can contain "pure" pixels when *ξ* = 1. We also introduce a sparsity parameter *ι* (*ι* > 0), which controls the sparsity of **S**. If *ι* is set at 0.8, 20% of the *J* × *I* elements in **S** are selected randomly and set to zeros at first, and then the nonzero elements in each column vector of **S** are generated following the Dirichlet pdf with the STU constraint and the maximal threshold *ξ*.

**N**, assumed to be zero-mean white Gaussian. The noise is characterized by the SNR

where *σ*^{2} is its variance.

*J*, the randomly selected endmember spectra,

*I, ξ, ι*, and the SNR. The default configuration is given in Table 1.

Default configuration of the synthetic data

Parameter | Default value |
---|---|

| 4 |

| 1000 |

| 0.8 |

| 0.8 |

SNR | Inf |

### 4.3. Compared algorithms

- (1)
F1-NMF: the basic HALS-NMF with no extra constraint based on HALS optimization algorithm. Only the nonnegative constraints are guaranteed.

- (2)
F2-NMF: the HALS-NMF is improved with the STU constraint.

- (3)
F3-NMF: the HALS-NMF with the STU and maximum spatial dispersion constraints.

- (4)
F4-NMF: the HALS-NMF with the STU and minimum spectral dispersion constraints.

- (5)
F5-NMF: the HALS-NMF with the STU and minimum distance constraints.

- (6)
F35-NMF: the HALS-NMF with the combined constraints of F3 and F5.

- (7)
VCA: a popular geometrical algorithm proposed in [4].

- (8)
BPSS2: an improved Bayesian algorithm addressed in [9] under nonnegativity and full additivity constraints.

- (9)
MiniDisCo: a novel NMF-based algorithm with spectral constraint given in [16].

Note that the initializations of **A** and **S** for all the algorithms are chosen from a uniform distribution on the interval [0,1].

### 4.4. Simulations

*f*function along the interactions of two optimization algorithms. Experiments 2-7 present statistical simulations to compare the average behaviors of the five F-NMF algorithms while varying the parameters given in Table 1, and robustness to an overestimation of the endmember number

*J*.

- (1)
The first experiment is to assess the choice of the optimization algorithm. We compare the convergence efficiency between the PG, which is widely used for NMF optimization, and the HALS algorithm. Here, the PG and HALS algorithms are regularized with the minimum spectral dispersion constraint. The PG-based algorithm is named MiniDisCo in [11], and HALS-based algorithm in this experiment is also called F4-NMF as above. The

*f*value is calculated with the corresponding constraints and the performances of the two estimators are presented in Figure 1. Note that both the curves result from the same HSI, with the same random initial conditions; thus, the only variability is the optimization method. We note that the final value of*f*is almost the same with both algorithms, whereas the convergence speed of HALS is faster.

_{1}= 1, α

_{2}= 0.1, β

_{1}= 0.1, and β

_{2}= 0.1, which are chosen to minimize the average evaluation errors for synthetic data.

- (2)
In the second experiment, the algorithms are compared when the number of endmembers

*J*varies.

*J*is set from 3 to 10 as the experiment in [11]. The performance metrics of SME are shown in Figure 2. Note that the considered statistics do not necessarily include each of the results. Here, SME values higher than 0.5 are not included. In particular, the PG-NMF results are never considered while the F-NMF results are all included, because the SME values of PG-NMF are always greater than 0.5. With NMF algorithms, only a local minimum can be attained in general. In the case of random initializations and no constraints, HALS is able to obtain a better solution than PG.

*J*higher to 20 to test the performance of F-NMF. The performance metrics are shown in Figure 3. Note that the F1-NMF without constraints performs worse as the number of endmembers increases. In the case of the constrained F-NMF (F2, F3, F4, F5, F35), the results are much better. Figure 3 puts forward the high robustness of the constrained F-NMF algorithms, when the basic F-NMF is sensitive to the number of endmembers. The combination of constraints F3 and F5, F35, gives good results.

- (3)
The purity level

*ξ*is the topic of the third experiment. None of the considered algorithms are based on the hypothesis of one pure pixel for each endmember, but the un-mixing performance may vary with the purity. The obtained performance metrics are presented in Figure 4. F3-NMF is particularly worse for AME when*ξ*= 0.6, because the low purity level make the maximum spatial dispersion constraint ineffective. F35-NMF also performs worse in term of AME due to the maximum spatial dispersion constraint.

*ξ*. The two NMF-based algorithms and BPSS2 are each initialized with VCA. The parameter of the spectral constraint is 0.1 for MiniDisCo. VCA performs better with higher purity level due to its assumption of pure pixels. MiniDisCo and F35-NMF both improve the unmixing results of VCA. Specifically, MiniDisCo and BPSS2 outperform F35-NMF in the sense of AME but the result is quite the reverse in the sense of SME, which is caused by different constraints in MiniDisCo and F35-NMF. In the sense of AME, F35-NMF performs worse as the purity level decreases, because the algorithm is regularized by the maximum spatial dispersion constraint, which improves the values of AME for the mixing data with high purity level. This could be verified by the results shown in Figure 4a. The algorithms with the maximum spatial dispersion constraint (F3-NMF and F35-NMF) give worse results than the other algorithms (F4-NMF and F5-NMF). We choose F35-NMF for comparison due to its better performances in SME and SAD. In the sense of SAD, MiniDisCo is better than F35-NMF with lower purity level, but F35-NMF performs better with purity level lower than 0.7. The performance of BPSS2 is always worse. This may be resulted by the minimum distance constraint in F35-NMF, which plays an important role in the unmixing of highly-mixed data.

- (4)
The fourth experiment studies the robustness to noise of the considered algorithms. The metric values obtained for various SNR are shown in Figure 6. The F-NMF algorithms are all based on the RQE minimization, which is optimal for white Gaussian noise. Thus, the performances do not significantly depend on the noise. In accordance with the experiment 3, the F3-NMF and F35-NMF results are not good in AME, but better in the terms of SME and SAD.

- (5)
It is interesting to study the estimation quality in terms of the data spatial dimensions. Figure 7 presents the influence of the number of observed spectral pixels. The F-NMF algorithms are both robust to a small number of spectral pixels and a large amount of data. It is interesting to see that a small number of spectral pixels globally improve the performances of the regularized NMF. The F4-NMF and F5-NMF outperform the other algorithms in AME, but the results of F3-NMF and F35-NMF algorithms are better in the terms of SME and SAD. In general, a large data set does not improve the results, so it is more efficient to use a small set of data (400 pixels).

- (6)
This experiment tests the influence of the sparsity parameter

*ι*. The results are presented in Figure 8. All the algorithms are not very sensitive to the sparsity parameter. The F4-NMF and F5-NMF outperform the other algorithms in AME, and the maximum spatial dispersion constraint brings improvement in SME and SAD.

- (7)
Estimating the endmember number

*J*is the first issue of the HSI analysis. On real data, existing methods to estimate*J*generally overestimate the number [22]. Thus, we study the robustness of the algorithms to an overestimation of*J*(Figure 9). Here, we overestimate*J*by 1. The estimation errors show that constrained F-NMF algorithms are robust to an overestimation of*J*, while the basic F-NMF is sensitive to the number of endmembers.

- (1)
The optimized algorithm of HALS outperforms PG in convergence speed and efficiency. In [11], poor estimations due to local minimum affect the basic PG-NMF, so the estimated values of SME higher than 0.5 are not included in the statistics. In F-NMF, the estimation performance is much better so all the results of the experiments are considered.

- (2)
The performances of constrained F-NMF are better than the basic NMF, according to all the parameters (

*J, ξ*, SNR,*I*, and*ι*) and the different performance metrics. - (3)
The NMF algorithms with minimum spectral dispersion constraint (F4-NMF) and minimum distance constraint (F5-NMF) performs better in AME, while the algorithms with maximum spatial dispersion constraint (F3-NMF and F35-NMF) outperform the other algorithms in the terms of SME and SAD.

- (4)
The NMF algorithm with combined constraints (F35-NMF) performs better than the algorithm with one constraint (F3-NMF).

- (5)
F-NMF algorithm can improve the unmixing results initialized by VCA.

## 5. Application on real hyperspectral data

It is required to estimate the number of endmembers *J* before unmixing the image. In this article, the number of endmembers is determined from the final RQE obtained after convergence for many preliminary experiences, and is set to *J* = 11; however, this value is only an approximation.

*J, I, ξ, ι*, SNR), when the endmembers are selected randomly from the library. We should obtain the same 11 identified references in each experiment. However, the results vary in the 50 experiments. In order to compare the results with a minimum volume-based algorithm, we choose SISAL for its good performances on simulated data and its high speed. Note that the F-NMF and SISAL algorithms are all based on the assumption that the endmembers, or at least some of them, are not in the data set. The references identified by F-NMF are presented in Tables 2, 3, 4, 5, 6, and 7 and the results by SISAL in Table 8. The estimated endmembers are identified as the closest library spectra in the sense of SAD. It can be seen from the tables that F3-NMF gives 77 names for a total of 550 possible different answers, whereas the other four F-NMFs give much more references. The top 11 responses of F3-NMF and F35-NMF represent 66.7 and 68%, respectively, of all the answers. All these results show the stability of F3-NMF, due to the maximum spatial dispersion constraint. From Table 8, we can see that the SISAL identifies 146 names from 550 possible answers, which shows its serious instability. Therefore, the F-NMF algorithms are more stable than SISAL. Otherwise, the mean SAD between the estimated endmembers and the closest references in the library is significantly lower with F-NMF than SISAL, so we can conclude that the F-NMF algorithms are more efficient in endmember identification for difficult real cases.

References identified by f1-NMF

USGS reference name | Percent (%) | SAD |
---|---|---|

Kaolin/Smect KLF508 85%K | 21.273 | 5.0403 |

Andradite HS111.3B | 7.6364 | 4.8941 |

Andradite WS488 | 3.8182 | 4.5956 |

Lepidolite NMNH105543 | 3.4545 | 6.9935 |

Goethite WS219 (limonite) | 3.0909 | 6.8237 |

Sphene HS189.3B | 3.0909 | 4.3982 |

Kaolin/Smect H89-FR-5 30K | 3.0909 | 3.8384 |

Barite HS79.3B | 2.5455 | 4.0984 |

Rectorite RAr-1 | 2.5455 | 3.5563 |

Kaolin/Smect KLF511 12%K | 2.3636 | 3.1599 |

Almandine WS477 | 2.3636 | 5.8007 |

Sum of the top 11 | 55.273 | |

83 names | 100 | 5.1017 |

References identified by f2-NMF

USGS reference name | Percent (%) | SAD |
---|---|---|

Kaolin/Smect KLF508 85%K | 15.455 | 5.1444 |

Andradite HS111.3B | 6.7273 | 4.5232 |

Andalusite NMNHR17898 | 5.4545 | 3.5174 |

Richterite HS336.3B | 3.6364 | 3.6131 |

Andradite WS487 | 3.2727 | 3.8255 |

Kaolin/Smect H89-FR-5 30K | 3.0909 | 3.7755 |

Kaolin/Smect KLF511 12%K | 2.5455 | 3.3961 |

Lepidolite NMNH105543 | 2.3636 | 7.2303 |

Rectorite RAr-1 | 2.3636 | 3.5788 |

Nontronite NG-1.a | 2.3636 | 4.3183 |

Goethite WS219 (limonite) | 2.1818 | 6.8177 |

Sum of the top 11 | 49.455 | |

98 names | 100 | 5.9825 |

References identified by f3-NMF

USGS reference name | Percent (%) | SAD |
---|---|---|

Kaolin/Smect KLF508 85%K | 26.545 | 4.5345 |

Andradite HS111.3B | 9.0909 | 4.2029 |

Andradite WS487 | 6 | 3.6531 |

Andalusite NMNHR17898 | 5.0909 | 3.6367 |

Richterite HS336.3B | 4.1818 | 3.2513 |

Kaolin/Smect H89-FR-5 30K | 3.8182 | 3.295 |

Illite IL105 (1Md) | 2.9091 | 3.8568 |

Montmorillonite+Illi CM37 | 2.5455 | 3.1543 |

Barite HS79.3B | 2.5455 | 3.2624 |

Kaolin/Smect KLF511 12%K | 2.1818 | 2.719 |

Rectorite RAr-1 | 1.8182 | 3.877 |

Sum of the top 11 | 66.727 | |

77 names | 100 | 5.1216 |

References identified by F4-NMF

USGS reference name | Percent (%) | SAD |
---|---|---|

Kaolin/Smect KLF508 85%K | 16.364 | 5.1932 |

Andradite HS111.3B | 6.7273 | 4.513 |

Andalusite NMNHR17898 | 5.8182 | 3.5355 |

Richterite HS336.3B | 4 | 3.674 |

Kaolin/Smect H89-FR-5 30K | 3.4545 | 3.829 |

Andradite WS487 | 3.2727 | 3.8056 |

Kaolin/Smect KLF511 12%K | 2.7273 | 3.3183 |

Lepidolite NMNH105543 | 2.5455 | 7.6882 |

Rectorite RAr-1 | 2.3636 | 3.518 |

Goethite WS219 (limonite) | 2.1818 | 6.8385 |

Nontronite NG-1.a | 2.1818 | 4.2891 |

Sum of the top 11 | 51.636 | |

97 names | 100 | 5.1216 |

References identified by F5-NMF

USGS reference name | Percent (%) | SAD |
---|---|---|

Kaolin/Smect KLF508 85%K | 16 | 5.1622 |

Andradite HS111.3B | 6.7273 | 4.5837 |

Andalusite NMNHR17898 | 5.8182 | 3.524 |

Richterite HS336.3B | 4.1818 | 3.6714 |

Andradite WS487 | 3.2727 | 3.806 |

Kaolin/Smect H89-FR-5 30K | 3.2727 | 3.7812 |

Lepidolite NMNH105543 | 2.5455 | 7.4252 |

Rectorite RAr-1 | 2.5455 | 3.6228 |

Nontronite NG-1.a | 2.3636 | 4.3196 |

Kaolin/Smect KLF511 12%K | 2.3636 | 3.3328 |

Goethite WS219 (limonite) | 2.1818 | 6.828 |

Sum of the top 11 | 51.273 | |

97 names | 100 | 5.8524 |

References identified by F35-NMF

USGS reference name | Percent (%) | SAD |
---|---|---|

Kaolin/Smect KLF508 85%K | 26.545 | 4.437 |

Andradite HS111.3B | 11.818 | 4.3172 |

Andradite WS487 | 6.7273 | 3.6406 |

Andalusite NMNHR17898 | 5.6364 | 3.6947 |

Kaolin/Smect KLF511 12%K | 3.2727 | 2.8078 |

Montmorillonite+Illi CM37 | 2.9091 | 3.2787 |

Richterite HS336.3B | 2.7273 | 3.4012 |

Illite IL105 (1Md) | 2.3636 | 3.784 |

Barite HS79.3B | 2 | 3.0921 |

Perthite HS415.3B | 2 | 2.774 |

Sphene HS189.3B | 2 | 3.6569 |

Sum of the top 11 | 68 | |

80 names | 100 | 4.9158 |

References identified by SISAL

USGS reference name | Percent (%) | SAD |
---|---|---|

Kaolin/Smect KLF508 85%K | 5.8182 | 7.1877 |

Acmite NMNH133746 | 4.9091 | 87.33 |

Hornblende_Fe HS115.3B | 4 | 24.591 |

Desert_Varnish GDS141 | 3.6364 | 6.5823 |

Limonite HS41.3 | 3.4545 | 29.357 |

Hematite FE2602 | 3.4545 | 4.6658 |

Lepidolite NMNH105543 | 3.2727 | 8.6374 |

Mordenite+Clinopt. GDS151 | 2.7273 | 3.1797 |

Almandine WS477 | 2.3636 | 5.2999 |

Rutile HS137.3B | 2.1818 | 5.5928 |

Almandine HS114.3B | 2.1818 | 27.247 |

Sum of the top 11 | 38 | |

146 names | 100 | 15.808 |

*y*-coordinate tick (from

*j*= 1 to

*J*) corresponds to zero reflectance of the

*j*th endmember. The associated spectral endmembers are the closest library spectra (Figure 11b) in the sense of SAD. Note that 3, 7, 8, and 10 spectra are all identified as Kaolin/Smect KLF508 85%K, whose proportion is 26% the first in Table 7. This is the reason of the low identification dispersion of F35-NMF in 50 runs. The estimated abundance maps are given in Figure 12, where the maximum abundance value

*ξ*

_{ j }of each endmember

*j*is high due to the maximum spatial dispersion constraint.

- (1)
The analyzed Cuprite data are only a selection of the whole scene, which holds 18 endmembers; thus, the unmixing results are also incomplete.

- (2)
It is difficult to find the right spectra in the considered library with a huge amount of references (500). Some prior knowledge should be used to reduce the number of references before the comparison.

- (3)
We use a linear mixing model in this article, but the radiative transfer is always nonlinear in real scene [10].

- (4)
It is subjective to identify the endmembers with SAD. A more robust method for identification should make the decision jointly with several criteria. Moreover, the variability of real spectra has made their identification from library more difficult.

Computation time of F-NMF algorithms(s)

F1 | F2 | F3 | F4 | F5 | F35 |
---|---|---|---|---|---|

0.23 | 0.24 | 0.24 | 0.40 | 0.40 | 0.40 |

## 6. Conclusion

In this article, we have proposed an NMF-based hyperspectral unmixing algorithm with flexible constraints, including the STU constraint, the maximum spatial dispersion constraint, the minimum spectral dispersion constraint, and the minimum distance constraint. The optimization scheme is based on the HALS, whose convergence speed outperforms that of PG. The resulting algorithm, called F-NMF, is experimentally tested with different constraints. The estimation accuracy shows that the F-NMF works stably in all experiments, overcoming the estimation instability of PG-NMF. In particular, the F-NMF algorithms are robust to high number of endmembers, low SNR, low number of observed pixels and overestimation of the number of endmembers.

The F-NMF algorithms seem to be effective in the estimation of abundance maps, since they consider the STU and maximum spatial dispersion constraints. The identified references of real data by F-NMF seem more stable and reliable than geometrical method like SISAL. However, the identified results of real data are unsatisfied so the identification method needs further investigation to improve the results.

## Declarations

### Acknowledgements

The authors would like to thank S. Moussaoui for letting us do the benchmarks with the code of his BPSS algorithms.

## Authors’ Affiliations

## References

- Nascimento JMP, Dias JMB: Does independent component analysis play a role in unmixing hyperspectral data?
*IEEE Trans Geosci Remote Sens*2005, 43(1):175-187.View ArticleGoogle Scholar - Theiler J, Lavenier D, Harvey N, Perkins S, Szymanski J:
*Using blocks of skewers for faster computation of pixel purity index.**Volume 4132*. Proc. of the SPIE International Conference on Optical Science and Technology, Bellingham, WA; 2000:61-71.Google Scholar - ME Winter: N-findr: an algorithm for fast autonomous spectral end-member determination in hyperspectral data. In
*Proc SPIE Conf Imaging Spectrometry V*.*Volume 3753*. Denver, Colorado; 1999:266-275.View ArticleGoogle Scholar - Nascimento J, Dias J: Vertex component analysis: a fast algorithm to unmix hyperspectral data.
*IEEE Trans Geosci Remote Sens*2005, 43(4):898-908.View ArticleGoogle Scholar - Li J, Bioucas-Dias JM: Minimum volume simplex analysis: a fast algorithm to unmix hyperspectral data.
*Proc IEEE IGARSS, Boston*2008, 3: 250-253.Google Scholar - Chan TH, Chi CY, Huang YM, Ma WK: A convex analysis-based minimum-volume enclosing simplex algorithm for hyperspectral unmixing.
*IEEE Trans Signal Process*2009, 57(11):4418-4432.MathSciNetView ArticleGoogle Scholar - Bioucas-Dias JM: A variable splitting augmented Lagrangian approach to linear spectral unmixing. In
*Proc WHISPERS*.*Volume 1*. Grenoble, France; 2009:1-4.Google Scholar - Moussaoui S, Brie D, Mohammad-Djafari A, Carteret C: Separation of non-negative mixture of non-negative sources using a Bayesian approach and MCMC sampling.
*IEEE Trans Signal Process*2006, 54(11):4133-4145.View ArticleGoogle Scholar - Dobigeon N, Moussaoui S, Tourneret J-Y, Carteret C: Bayesian separation of spectral sources under non-negativity and full additivity constraints.
*Signal Process*2009, 89(12):2657-2669. 10.1016/j.sigpro.2009.05.005View ArticleMATHGoogle Scholar - Schmidt F, Schmidt A, Tréguier E, Guiheneuf M, Moussaoui S, Dobigeon N: Implementation strategies for hyperspectral unmixing using Bayesian source separation.
*IEEE Trans Geosci Remote Sens*2010, 48(11):4003-4013.Google Scholar - Lee DD, Seung HS: Learning the parts of objects by non-negative matrix factorization.
*Nature*1999, 401(6755):788-791. 10.1038/44565View ArticleGoogle Scholar - Heinz DC, Chang CI: Fully constrained least squares linear spectral mixture analysis method for material quantification in hyperspectralimagery.
*IEEE Trans Geosci Remote Sens*2001, 39(3):529-545. 10.1109/36.911111View ArticleGoogle Scholar - Jia S, Qian Y: Constrained nonnegative matrix factorization for hyperspectral unmixing.
*IEEE Trans Geosci Remote Sens*2009, 47(1):161-173.View ArticleMATHGoogle Scholar - Miao L, Qi H: Endmember extraction from highly mixed data using minimum volume constrained nonnegative matrix factorization.
*IEEE Trans Geosci Remote Sens*2007, 45(3):765-777.View ArticleGoogle Scholar - Yu Y, Guo S, Sun WD: Minimum distance constrained nonnegative matrix factorization for the endmember extraction of hyperspectral images.
*Proc SPIE, MIPPR, Wuhan, China*2007, 6790: 1-4.Google Scholar - Huck A, Guillaume M, Blanc-Talon J: Minimum dispersion constrained nonnegative matrix factorization to unmix hyperspectral data.
*IEEE Trans Geosci Remote Sens*2010, 48(6):2590-2602.View ArticleGoogle Scholar - Huck A, Guillaume M: Robust hyperspectral data unmixing with spatial and spectral regularized NMF.
*Proc WHISPERS, Reykjavik, Iceland*2010, 2: 1-4.Google Scholar - Cichocki A, Phan AH, Caiafa C: Flexible HALS algorithms for sparse non-negative matrix/tensor factorization.
*IEEE Workshop on Machine Learning for Signal Processing, Cancun, Mexico*2008, 4(4):73-78.Google Scholar - Gillis N, Glineur F: Using underapproximations for sparse nonnegative matrix factorization.
*Pattern Recogn*2010, 43(4):1676-1687. 10.1016/j.patcog.2009.11.013View ArticleMATHGoogle Scholar - Gillis N, Glineur F: Nonnegative factorization and the maximum edge biclique problem. 2008.Google Scholar
- Ho N-D: Nonnegative matrix factorization-algorithms and applications.
*PhD thesis*. Universite catholique de Louvain; 2008.Google Scholar - Chang C-I, Du Q: Estimation of number of spectrally distinct signal sources in hyperspectral imagery.
*IEEE Trans Geosci Remote Sens*2004, 42(3):608-619. 10.1109/TGRS.2003.819189View ArticleGoogle Scholar - [http://speclab.cr.usgs.gov/PAPERS/cuprite.gr.truth.1992/swayze.1992.html]

## Copyright

This article is published under license to BioMed Central Ltd. This is an Open Access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/2.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.