Introduction
Hyperspectral (HS) imaging has attracted wide attention in recent years since it can simultaneously obtain images of the same scenario across plenty of different successive wavelengths at the same time [1]–[3]. Because hyperspectral image (HSI) has rich spectral information, it has been widely used in many fields, such as target detection [4], environmental monitoring [5], military [6], and remote sensing [7]. However, since there is a limited amount of incident energy in optical remote sensing systems, the imaging systems have to compromise between the spectral resolution and spatial resolution [8]. For example, HSIs captured by HYPXIM usually have more than one hundred spectral bands but only a decametric spatial resolution. Compared with HS imaging sensors, the multispectral (MS) imaging sensors can provide multispectral images (MSIs) with much higher spatial resolution but with a limited number of spectral bands. For example, the PLEIADES can provide MSIs with a spatial resolution of 70 cm but with only three or four spectral bands. In order to enhance the spatial resolution of HSIs, researchers have made much effort. A popular approach to reconstruct the high spatial resolution HSI (HR-HSI) is to fuse the high spatial resolution MSI (HR-MSI) with the low spatial resolution HSI (LR-HSI) [9], [10]. This approach is called HSI–MSI fusion or HSI super-resolution.
HSI super-resolution problem aims to reconstruct an HR-HSI by fusing the spectral information of an LR-HSI and the spatial information of an HR-MSI, as illustrated in Fig. 1. Note that the LR-HSI and the HR-MSI should be the same scene. The target HSI should not only have a good visual effect but also ensure the authenticity of each pixel.
A large number of studies have been done on HSI super-resolution. A special situation of HSI super-resolution is pansharpening, which fuses an LR-HSI with its corresponding panchromatic (PAN) image [9], [11]. A variety of pansharpening methods have been proposed over the past two decades. Generally, these methods can be categorized into two classes, i.e., transform-based methods [12]–[14] and variational methods [15]–[17]. However, because the PAN images have little spectral resolution, there are usually considerable spectral distortions in the HR-HSIs reconstructed by these pansharpening methods.
As the MSIs contain more spectral information than the PAN images, in recent work, HSI–MSI fusion, which can be seen as the extension of pansharpening, has drawn more attention. Yokoya et al. [10] present a comparative review of several HSI–MSI fusion techniques. Typically, the HSI–MSI fusion methods can be divided into four categories: component substitution (CS), Bayesian, deep learning, and sparse representation.
In the CS-based approaches, a basic idea is to substitute one component of the HSI with the high-resolution (HR) image. For example, the intensity-hue-saturation (IHS) [18], [19] method replaces the intensity component in the IHS domain of the LR image with the PAN image. The principal component analysis [20] method uses the HR image to replace the first principal component of the LR-HSI. However, the CS-based approaches usually result in spectral distortions in the obtained HR-HSI.
The Bayesian-based approaches introduce the appropriate prior distribution of the HR-MSI, such as naive Gaussian [21], [22] and sparsity promoting prior [23], [24] to achieve the accurate estimation. The variational methods can be regarded as a special case of the Bayesian one. The target images are estimated by minimizing the objective function, which is structured by the posterior probability density of the fused image. Among these methods, HS super-resolution [25] uses the vector-total-variation-based regularization in the objective function. Zhang et al. [26], [27] introduced a method that works in the wavelet domain and later published an expectation–maximization algorithm to maximize the posterior distribution.
Since the deep learning has been demonstrated to be very effective in object detection [28]–[30], classification [31]–[33], and natural image super-resolution [34]–[36], many researchers have introduced deep learning into HSI super-resolution. Li et al. [37] proposed to learn an end-to-end spectral difference mapping between the LR-HSI and HR-HSI through a deep spectral difference convolution neural network. Yuan et al. [38] proposed a multiscale and multidepth convolution neural network to achieve the HR-HSI. In order to take the advantage of the spectral correlation and exploit the HR-MSI, Yang et al. [39] presented a convolution neural network with two branches. With the two branches convolution neural network, the spectrum features of each pixel and its corresponding spatial neighborhood are extracted from the LR-HSI and the HR-MSI, respectively. Dian et al. [40] proposed to learn the spectral prior of HSI via deep residual convolutional neural networks. In addition to the convolution neural network, a stacking sparse denoising autoencoder-based deep neural network is proposed by Huang et al. [41] for pansharpening. Although the deep learning based methods obtained great reconstruction results, these kinds of methods need large amounts of training samples to estimate the parameters.
In the past years, the sparse representation has been widely used in remote sensing applications [42]. The sparse representation-based HSI super-resolution methods usually represent the targeted HR-HSI image by the product of a spectral basis matrix and a coefficient matrix, where the spectral basis and coefficient matrices can be extracted from the LR-HSI and the HR-MSI. Besides, some matrix factorization and unmixing-based methods can also be regarded as the sparse representation-based method because the source images are decomposed into spectral bases and coefficients. Actually, the sparse representation-based methods are usually combined with matrix factorization and spectral unmixing. Based on the unsupervised spectral unmixing, Yokoya et al. [43] proposed a coupled nonnegative matrix factorization (CNMF) approach to estimate the HSI endmember matrix and the HR abundance matrix. However, the nonnegative matrix factorization is usually not unique. So, Yokoya et al. [43] cannot always obtain satisfactory results. Huang et al. [44] used the k-singular value decomposition (K-SVD) algorithm [45] to learn the spectral basis and proposed a sparse prior-based matrix factorization method to fuse the remote sensing MSI at different spatial and spectral resolution. Zhang et al. [46] used the group spectral embedding and low-rank factorization to fuse the LR-HSI and HR-MSI. Lanaras et al. [47] proposed to jointly solve the spectral unmixing problems for both input images. However, only using a spectral dictionary is insufficient for preserving spatial information, and vice-versa. To address this problem, an HSI–MSI fusion method termed optimized twin dictionaries (OTD) using optimized twin dictionaries was proposed by Han et al. [48]. Since the pixelwise sparse representation neglects the similarity among neighbor pixels, Akhtar et al. [49] proposed to utilize the similarities among the spectral pixels in the same local patch and obtain the coefficients with a generalization of simultaneous orthogonal matching pursuit (G-SOMP+) algorithm for each local patch. Later, Akhtar et al. [50] proposed a Bayesian dictionary learning and Bayesian sparse coding approach for HSI super-resolution and achieved improved performance. Note that, the structures of MSI are usually very complex, and thus, a fixed local window may still contain different variations. Combined with superpixel segmentation methods, Fang et al. [51] proposed a superpixel-based sparse representation (SSR) model, which ensured that the shape and size of each superpixel can adaptively adjust according to the spatial structures of MSI, and therefore, the spatial structures of spectral pixels in each superpixel are similar for HSI super-resolution. Furthermore, Dong et al. [52] proposed a nonnegative structured sparse representation (NSSR) method, which exploited a clustering-based structured sparse coding approach to ensure the spatial correlation among the obtained sparse coefficients.
Sparse representation-based approaches are indeed effective for HSI super-resolution and achieve great reconstruction results. However, the existing methods usually use an
The main contributions of this article can be summarized as follows.
We introduce the ASR, which can obtain more precise sparse coefficients by balancing the sparsity and correlation of the coefficients, into the HSI super-resolution model.
Instead of keeping the spectral basis fixed, we alternately optimize the spectral basis and sparse coefficients.
We design two specific ADMM methods to update the spectral basis and sparse coefficients, respectively.
Experimental results on both ground-based HSIs and real remote sensing HSIs show that our ANSR method performs better than some other state-of-the-art HSI super-resolution methods.
The remainder of this article is organized as follows. We briefly introduce the spectral dictionary learning method in [52] and the ASR in Section II. In Section III, we first formulate the problem of HSI super-resolution and then describe the details of the proposed ANSR method for HSI super-resolution. Extensive experiments and comparisons are shown in Section IV. Finally, Section V concludes this article.
Related Work
In this section, we introduce the spectral dictionary learning method in [52] and the ASR, which are used in our method.
A. Spectral Dictionary Learning
We denote the LR-HSI as \begin{equation*}
\boldsymbol{X}\ = \ \boldsymbol{D}\boldsymbol{B} + \boldsymbol{V} \tag{1}
\end{equation*}
In (1), both \begin{align*}
\left({\boldsymbol{D},\ \boldsymbol{B}} \right) =& \ {\rm{arg}}\mathop {\min }\limits_{\boldsymbol{D},\boldsymbol{B}} \frac{1}{2}\|\boldsymbol{X} - \boldsymbol{D}\boldsymbol{B}\|_F^2 + \lambda \|\boldsymbol{B}{\|_1} \\
{\rm{s}}.{\rm{t}}.\ \boldsymbol{B} \ge& 0,\ \boldsymbol{D} \geq 0 \tag{2}
\end{align*}
Because that the sparse coefficient matrix
With \begin{equation*}
\boldsymbol{B}\ = \ {\rm{arg}}\mathop {\min }\limits_{\boldsymbol{B}} \frac{1}{2}\|\boldsymbol{X} - \boldsymbol{D}\boldsymbol{B}\|_F^2 + \lambda \|\boldsymbol{B}{\|_1},\ {\rm{s}}.{\rm{t}}.\ \boldsymbol{B} \geq 0 \tag{3}
\end{equation*}
\begin{align*}
L\!\left({\boldsymbol{B},\ \boldsymbol{S},\ \boldsymbol{U}} \right) =& \frac{1}{2}\ \|\boldsymbol{X} \!-\! \boldsymbol{D}\boldsymbol{B}\|_F^2 \!+\! \lambda \|\boldsymbol{B}{\|_1} \!+\! \mu \|\boldsymbol{S}\! -\! \boldsymbol{B} \!+\! \frac{{\boldsymbol{U}}}{{2\mu }}\|_F^2 \\
{\rm{s}}.{\rm{t}}.\ \boldsymbol{B} \ge& 0 \tag{4}
\end{align*}
With \begin{equation*}
\boldsymbol{D}\ = \ {\rm{arg}}\mathop {\min }\limits_{\boldsymbol{D}} \|\boldsymbol{X} - \boldsymbol{D}\boldsymbol{B}\|_F^2,{\rm{\ s}}.{\rm{t}}.{\rm{\ }}\boldsymbol{D} \geq 0. \tag{5}
\end{equation*}
Similar to the online dictionary learning method, (5) is solved by using block coordinate descent. During each iteration, one column of
More information about the spectral dictionary learning method can be found in [52].
B. Adaptive Sparse Representation
As we all know, the goal of sparse representation is to encode a signal vector as a linear combination of a few dictionary atoms. Suppose that \begin{equation*}
\mathop {\min }\limits_{\boldsymbol{\alpha }} \|\boldsymbol{x} - \boldsymbol{D}_{\boldsymbol{s}}\boldsymbol{\alpha }\|_2^2 + \lambda \|\boldsymbol{\alpha }{\|_0} \tag{6}
\end{equation*}
However, the \begin{equation*}
\mathop {\min }\limits_{\boldsymbol{\alpha }} \|\boldsymbol{x} - \boldsymbol{D} {_{\boldsymbol{s}}}\boldsymbol{\alpha }\|_2^2 + \lambda \|\boldsymbol{\alpha }{\|_1}. \tag{7}
\end{equation*}
The result of (7) can be solved by quadratic programming techniques, including basis pursuit [56], LASSO [57], etc.
Although the \begin{equation*}
\mathop {\min }\limits_{\boldsymbol{\alpha }} \|\boldsymbol{x} - \boldsymbol{D}{_{\boldsymbol{s}}}\boldsymbol{\alpha }\|_2^2 + \lambda \|\boldsymbol{\alpha }\|_2^2. \tag{8}
\end{equation*}
In contrast, the collaborative representation model only takes the correlation into consideration and completely ignores the sparsity. Actually, the best choice is to balance the sparsity and correlation and make a compromise between \begin{equation*}
\|\boldsymbol{\alpha }{\|_2} \leq \|\boldsymbol{D}{_{\boldsymbol{s}}}{\rm{Diag}}\left({\boldsymbol{\alpha }} \right){\|_*} \leq \|\boldsymbol{\alpha }{\|_1} \tag{9}
\end{equation*}
\begin{equation*}
\mathop {\min }\limits_{\boldsymbol{\alpha }} \|\boldsymbol{x} - \boldsymbol{D}{_{\boldsymbol{s}}}\boldsymbol{\alpha }\|_2^2 + \lambda \|\boldsymbol{D}{_{\boldsymbol{s}}}{\rm{Diag}}\left({\boldsymbol{\alpha }} \right){\|_*}. \tag{10}
\end{equation*}
Based on the convexity of the model, this problem can be solved by some efficient methods. Among them, the alternation direction method of multipliers (ADMM) [62], [63] is a widely used method to find an approximate optimal solution.
Proposed ANSR Method
In this section, we first provide a general introduction of the HSI super-resolution problem, including the linear spectral mixture model. Then, we introduce our super-resolution model in detail. Finally, we brief readers on the alternating optimization method thoroughly, including the optimization of coefficients and spectral basis.
In this article, the bold lowercase letters stand for the vectors and the bold uppercase letters stand for the matrices. The plain lowercase letters stand for the scalars.
A. Problem Formulation
The HSI super-resolution aims to recover an HR-HSI
In the linear spectral mixture model, each spectral vector \begin{equation*}
\boldsymbol{z}{_{\boldsymbol{i}}} = \ \boldsymbol{D}\boldsymbol{\alpha }{_{\boldsymbol{i}}} \tag{11}
\end{equation*}
\begin{equation*}
\boldsymbol{Z}\ = \ \boldsymbol{D}\boldsymbol{A} \tag{12}
\end{equation*}
Furthermore, both \begin{equation*}
\boldsymbol{X}\ = \ \boldsymbol{Z}\boldsymbol{H} \tag{13}
\end{equation*}
The HR-MSI \begin{equation*}
\boldsymbol{Y}\ = \ \boldsymbol{P}\boldsymbol{Z} \tag{14}
\end{equation*}
By combining the linear mixture model (12) and the forward models (13) and (14), we have
\begin{align*}
\boldsymbol{X}=& \boldsymbol{Z}\boldsymbol{H}\ = \ \boldsymbol{D}\boldsymbol{A}\boldsymbol{H}\ = \ \boldsymbol{D}\boldsymbol{B} \tag{15}\\
\boldsymbol{Y}=& \boldsymbol{P}\boldsymbol{Z}\ = \ \boldsymbol{P}\boldsymbol{D}\boldsymbol{A} \tag{16}
\end{align*}
According to the linear spectral mixture model (12), the HSI super-resolution problem can be transformed into the estimation of spectral basis
B. Establishment of Our Model
As mentioned above, the HSI super-resolution problem can be transformed into the estimation of spectral basis \begin{equation*}
\mathop {\min }\limits_{\boldsymbol{D},\ \boldsymbol{A}} \|\boldsymbol{Y} - \boldsymbol{P}\boldsymbol{D}\boldsymbol{A}\|_F^2 + \|\boldsymbol{X} - \boldsymbol{D}\boldsymbol{A}\boldsymbol{H}\|_F^2. \tag{17}
\end{equation*}
Obviously, the above optimization problem is ill-posed, and the solutions of
The sparsity prior is known to be a very effective method to deal with the HSI super-resolution problem. With the sparsity constraint, we assume that each spectral pixel in the target HSI can be represented as a linear combination of a few distinct atoms of spectral basis. Then, the HSI super-resolution problem can be written as
\begin{equation*}
\mathop {\min }\limits_{\boldsymbol{D},\ \boldsymbol{A}} \|\boldsymbol{Y} - \boldsymbol{P}\boldsymbol{D}\boldsymbol{A}\|_F^2 + \|\boldsymbol{X} - \boldsymbol{D}\boldsymbol{A}\boldsymbol{H}\|_F^2 + \eta \|\boldsymbol{A}{\|_1} \tag{18}
\end{equation*}
However, in (18), the sparse coefficients of each spectral pixel are estimated independently. It is generally known that a pixel of a typical HSI usually has a strong spatial correlation with its similar neighbors. In order to take advantage of the local and nonlocal similarities, we assume that a spectral pixel \begin{align*}
& \mathop {\min }\limits_{\boldsymbol{D},\ \boldsymbol{A}} \|\boldsymbol{Y} - \boldsymbol{P}\boldsymbol{D}\boldsymbol{A}\|_F^2 + \|\boldsymbol{X} - \boldsymbol{D}\boldsymbol{A}\boldsymbol{H}\|_F^2 + {\eta _2}\|\boldsymbol{A}{\|_1} \\
&\quad + {\eta _1}\mathop \sum \limits_{q\ = \ 1}^Q \mathop \sum \limits_{i \in {S_q}} \|\boldsymbol{D}\boldsymbol{\alpha }{_{\boldsymbol{i}}} - \boldsymbol{\mu }{_{\boldsymbol{q}}}\|_2^2 \tag{19}
\end{align*}
\begin{equation*}
\boldsymbol{\mu }{_{\boldsymbol{q}}} = \mathop \sum \limits_{i \in {S_q}} \boldsymbol{\omega }{_{\boldsymbol{i}}}\left({\boldsymbol{D}\boldsymbol{\alpha }{_{\boldsymbol{i}}}} \right)\ \tag{20}
\end{equation*}
\begin{equation*}
\boldsymbol{\omega }{_{\boldsymbol{i}}} = \frac{1}{c}{\rm{\ exp}}\left({\frac{{ - \|\boldsymbol{y}{_{\boldsymbol{i}}} - \boldsymbol{y}{_{\boldsymbol{q}}}\|_2^2}}{h}} \right) \tag{21}
\end{equation*}
\begin{align*}
& \mathop {\min }\limits_{\boldsymbol{D},\ \boldsymbol{A}} \|\boldsymbol{Y} - \boldsymbol{P}\boldsymbol{D}\boldsymbol{A}\|_F^2 + \|\boldsymbol{X} - \boldsymbol{D}\boldsymbol{A}\boldsymbol{H}\|_F^2 + {\eta _2}\|\boldsymbol{A}{\|_1} \\
& + {\eta _1}\|\boldsymbol{D}\boldsymbol{A} - \boldsymbol{U}\|_F^2 \tag{22}
\end{align*}
Besides, considering the physical characteristics of HSIs, the pixels of an HSI should be nonnegative. With this nonnegative prior, we can improve (22) to
\begin{align*}
& \mathop {\min }\limits_{\boldsymbol{D},\ \boldsymbol{A}} \|\boldsymbol{Y} - \boldsymbol{P}\boldsymbol{D}\boldsymbol{A}\|_F^2 + \|\boldsymbol{X} - \boldsymbol{D}\boldsymbol{A}\boldsymbol{H}\|_F^2 + {\eta _2}\|\boldsymbol{A}{\|_1} \\
& \quad + {\eta _1}\|\boldsymbol{D}\boldsymbol{A} - \boldsymbol{U}\|_F^2,\ {\rm{s}}.{\rm{t}}.\ \boldsymbol{A} \geq 0,\ 0 \leq \boldsymbol{D} \leq 1. \tag{23}
\end{align*}
Furthermore, in order to balance the sparsity and correlation, we propose to use trace LASSO instead of the \begin{align*}
&\mathop {\min }\limits_{\boldsymbol{D},\ \boldsymbol{A}} \|\boldsymbol{Y} - \boldsymbol{P}\boldsymbol{D}\boldsymbol{A}\|_F^2 + \|\boldsymbol{X} - \boldsymbol{D}\boldsymbol{A}\boldsymbol{H}\|_F^2 + {\eta _1}\|\boldsymbol{D}\boldsymbol{A} - \boldsymbol{U}\|_F^2 \\
& + {\eta _2}\mathop \sum \limits_{i\ = \ 1}^N \|\boldsymbol{P}\boldsymbol{D}{\rm{Diag}}\left({\boldsymbol{\alpha }{_{\boldsymbol{i}}}} \right){\|_*}\\
&{\rm{s}}.{\rm{t}}.\ \boldsymbol{A} \geq 0,\ 0 \leq \boldsymbol{D} \leq 1 \tag{24}
\end{align*}
Once we have solved
C. Alternating Optimization of the Fusion Problem
It is obvious that (24) is highly nonconvex. However, the problem (24) is convex with respect to
Algorithm 1 : ANSR-Based HSI Super-Resolution.
Input: LR-HSI
Initialize the spectral basis
While not converge do
Update coefficient matrix
Update spectral basis
End while
Compute the desired HR-HSI
Output: HR-HSI
D. Optimization of the Coefficients With the Spectral Basis Fixed
In this procedure, we fix the spectral basis \begin{align*}
& \mathop {\min }\limits_{\boldsymbol{A}} \|\boldsymbol{Y} - \boldsymbol{P}\boldsymbol{D}\boldsymbol{A}\|_F^2 + \|\boldsymbol{X} - \boldsymbol{D}\boldsymbol{A}\boldsymbol{H}\|_F^2 + {\eta _1}\|\boldsymbol{D}\boldsymbol{A} - \boldsymbol{U}\|_F^2 \\
&\quad + {\eta _2}\mathop \sum \limits_{i\ = \ 1}^N \|\boldsymbol{P}\boldsymbol{D}{\rm{Diag}}\left({\boldsymbol{\alpha }{_{\boldsymbol{i}}}} \right){\|_*},\ {\rm{s}}.{\rm{t}}.\ \boldsymbol{A} \geq 0 \tag{25}
\end{align*}
Obviously, the optimization problem (25) is convex and can be efficiently solved by ADMM, which can decompose the complex optimization problem into several easily solved subproblems. In specific, we introduce \begin{align*}
& L\ \left({\boldsymbol{A},\ \boldsymbol{S},\ \boldsymbol{Q},\boldsymbol{Z},\ \boldsymbol{V}{_{\boldsymbol{1}}},\ \boldsymbol{V}{_{\boldsymbol{2}}},\ \boldsymbol{V}{_{\boldsymbol{3}}}} \right) \\
&\quad= \|\boldsymbol{Y} - \boldsymbol{P}\boldsymbol{D}\boldsymbol{S}\|_F^2\ + \|\boldsymbol{X} - \boldsymbol{Z}\boldsymbol{H}\|_F^2 \\
&\quad\quad + {\eta _1}\|\boldsymbol{D}\boldsymbol{S} \!-\! \boldsymbol{U}\|_F^2 \!+\! {\eta _2}\mathop \sum \limits_{i\ = \ 1}^N \|\boldsymbol{Q}{_{\boldsymbol{i}}}{\|_*} \!+\! \mu \|\boldsymbol{D}\boldsymbol{S} - \boldsymbol{Z} + \frac{{\boldsymbol{V}{_{\boldsymbol{1}}}}}{{2\mu }}\|_F^2\\
&\quad\quad\!+\! \mu \|\boldsymbol{S} \!-\! \boldsymbol{A} \!+\! \frac{{\boldsymbol{V}{_{\boldsymbol{2}}}}}{{2\mu }}\|_F^2 \!\!+\!\! \mu \mathop \sum \limits_{i\ = \ 1}^N \|\boldsymbol{Q}{_{\boldsymbol{i}}}\! \!-\!\! \boldsymbol{P}\boldsymbol{D}{\rm{Diag}}\left({\boldsymbol{\alpha }{_{\boldsymbol{i}}}} \right) \!+\! \frac{{\boldsymbol{V}_{\boldsymbol{3}}^{\left({\boldsymbol{i}} \right)}}}{{2\mu }}\|_F^2\\
&\quad\quad {\rm{s}}.{\rm{t}}.\ \boldsymbol{A} \geq 0 \tag{26}
\end{align*}
\begin{equation*}
\begin{array}{rcl} {\boldsymbol{S}^{\left({\boldsymbol{t} + 1} \right)}} = {\rm{arg}}\mathop {\min }\limits_{\boldsymbol{S}} L\left({{\boldsymbol{A}^{\left(\boldsymbol{t} \right)}},\boldsymbol{S},{\boldsymbol{Q}^{\left(\boldsymbol{t} \right)}},{\boldsymbol{Z}^{\left(\boldsymbol{t} \right)}},\boldsymbol{V}_1^{\left(\boldsymbol{t} \right)},\boldsymbol{V}_2^{\left(\boldsymbol{t} \right)},\boldsymbol{V}_3^{\left(\boldsymbol{t} \right)}} \right)\\
{\boldsymbol{Z}^{\left({\boldsymbol{t} + 1} \right)}} = {\rm{arg}}\mathop {\min }\limits_{\boldsymbol S} L\left({{\boldsymbol{A}^{\left(\boldsymbol{t} \right)}}\boldsymbol{,}{\boldsymbol{S}^{\left(\boldsymbol{t} \right)}}\boldsymbol{,}{\boldsymbol{Q}^{\left(\boldsymbol{t} \right)}}\boldsymbol{,Z,V}_1^{\left(\boldsymbol{t} \right)},\boldsymbol{V}_2^{\left(\boldsymbol{t} \right)},\boldsymbol{V}_3^{\left(\boldsymbol{t} \right)}} \right)\\ {\boldsymbol{Q}^{\left({\boldsymbol{t} + 1} \right)}} = {\rm{arg}}\mathop {\min }\limits_{\boldsymbol S} L\left({{\boldsymbol{A}^{\left(\boldsymbol{t} \right)}}\boldsymbol{,}{\boldsymbol{S}^{\left(\boldsymbol{t} \right)}}\boldsymbol{,Q,}{\boldsymbol{Z}^{\left(\boldsymbol{t} \right)}}\boldsymbol{,V}_1^{\left(\boldsymbol{t} \right)},\boldsymbol{V}_2^{\left(\boldsymbol{t} \right)},\boldsymbol{V}_3^{\left(\boldsymbol{t} \right)}} \right)\\ {\boldsymbol{A}^{\left({\boldsymbol{t} + 1} \right)}} = {\rm{arg}}\mathop {\min }\limits_{\boldsymbol S} L\left({\boldsymbol{A,}{\boldsymbol{S}^{\left(\boldsymbol{t} \right)}}\boldsymbol{,}{\boldsymbol{Q}^{\left(\boldsymbol{t} \right)}}\boldsymbol{,}{\boldsymbol{Z}^{\left(\boldsymbol{t} \right)}}\boldsymbol{,V}_1^{\left(\boldsymbol{t} \right)},\boldsymbol{V}_2^{\left(\boldsymbol{t} \right)},\boldsymbol{V}_3^{\left(\boldsymbol{t} \right)}} \right). \end{array} \tag{27}
\end{equation*}
Meanwhile, the Lagrangian multipliers are updated by
\begin{align*}
\boldsymbol{V}_1^{\left({\boldsymbol{t} + 1} \right)} =& \boldsymbol{V}_1^{\left(\boldsymbol{t} \right)}\ + \mu \left({\boldsymbol{D}{\boldsymbol{S}^{\left({\boldsymbol{t} + 1} \right)}} - {\boldsymbol{Z}^{\left({\boldsymbol{t} + 1} \right)}}} \right)\\
\boldsymbol{V}_2^{\left({\boldsymbol{t} + 1} \right)} =& \boldsymbol{V}_2^{\left(\boldsymbol{t} \right)}\ + \mu \left({{\boldsymbol{S}^{\left({\boldsymbol{t} + 1} \right)}} - {\boldsymbol{A}^{\left({\boldsymbol{t} + 1} \right)}}} \right)\\
\boldsymbol{V}_3^{\left(\boldsymbol{i} \right)\left({\boldsymbol{t} + 1} \right)} =& \boldsymbol{V}_3^{\left(\boldsymbol{i} \right)\left(\boldsymbol{t} \right)}\ + \mu \left({\boldsymbol{Q}_{\boldsymbol i}^{\left({\boldsymbol{t} + 1} \right)} - \boldsymbol{PDDiag}\left({\boldsymbol{\alpha }_{\boldsymbol i}^{\left({\boldsymbol{t} + 1} \right)}} \right)} \right).\tag{28}
\end{align*}
All the subproblems in (27) can be solved analytically, i.e.,
\begin{align*}
\boldsymbol{S}\ =& {\left[ {{{\left({\boldsymbol{PD}} \right)}^T}\left({\boldsymbol{PD}} \right) + \left({{\eta _1} + \mu } \right){\boldsymbol{D}^T}\boldsymbol{D} + \mu \boldsymbol{I}} \right]^{ - 1}}\\
&\bigg[ {{\left({\boldsymbol{PD}} \right)}^T}\boldsymbol{Y} + {\rm{\ }}{\eta _1}{\boldsymbol{D}^T}\boldsymbol{U} + \mu {\boldsymbol{D}^T}\left({\boldsymbol{Z} - \frac{{{\boldsymbol{V}_1}}}{{2\mu }}} \right) \\
&+ \mu \left({\boldsymbol{A} - \frac{{{\boldsymbol{V}_2}}}{{2\mu }}} \right) \bigg]\\
\boldsymbol{Z} =& \left[ {\boldsymbol{X}{\boldsymbol{H}^T} + \mu \left({\boldsymbol{DS} + \frac{{{\boldsymbol{V}_1}}}{{2\mu }}} \right)} \right]{\rm{\ }}{\left({\boldsymbol{H}{\boldsymbol{H}^T} + \mu \boldsymbol{I}} \right)^{ - 1}}
\end{align*}
\begin{align*}
{\boldsymbol{Q}_{\boldsymbol i}} =& {\mathcal{J}_{\frac{{{\eta _2}}}{{2\mu }}}}\ \left[ {\boldsymbol{PD}{\rm{Diag}}\left({{\boldsymbol{\alpha }_{\boldsymbol i}}} \right) - \frac{{\boldsymbol{V}_3^{\left(\boldsymbol{i} \right)}}}{{2\mu }}} \right]\\
{\boldsymbol{\alpha }_{\boldsymbol i}} =& \left\{ {{{\left[ {2\mu \boldsymbol{I} + 2\mu {\rm{Diag}}\left({{\rm{Diag}}\left({{{\left({\boldsymbol{PD}} \right)}^T}\left({\boldsymbol{PD}} \right)} \right)} \right)} \right]}^{ - 1}}} \right.\\
&\left[ 2\mu {\boldsymbol{s}_{\boldsymbol i}} + \ \boldsymbol{V}_2^{\boldsymbol i} + {\rm{Diag}} \left({{{\left({\boldsymbol{PD}} \right)}^T}\boldsymbol{V}_3^{\left(\boldsymbol{i} \right)}} \right)\right.\\
&\left. \left. + 2\mu {\rm{Diag}}\left({{{\left({\boldsymbol{PD}} \right)}^T}{\boldsymbol{Q}_{\boldsymbol i}}} \right) \right] \right\}_ {+} \tag{29}
\end{align*}
The update of variables and multipliers is alternately iterated until convergence. The overall algorithm for updating coefficient matrix
Algorithm 2 : Update \boldsymbol{A} With \boldsymbol{D} Fixed.
Input: LR-HSI
Initialization:
While not converge do
Update variables
Update Lagrangian multipliers
Update
Update
End while
Output: coefficient matrix
E. Optimization of the Spectral Basis With the Coefficients Fixed
In this procedure, we fix the coefficient matrix \begin{equation*}
\mathop {\min }\limits_{\boldsymbol{D}} \|\boldsymbol{Y} - \boldsymbol{P}\boldsymbol{D}\boldsymbol{A}\|_F^2 + \|\boldsymbol{X} - \boldsymbol{D}\boldsymbol{A}\boldsymbol{H}\|_F^2,\ {\rm{s}}.{\rm{t}}.\,0 \leq \boldsymbol{D} \leq 1. \tag{30}
\end{equation*}
As the nonlocal spatial similarity prior is mainly reflected by the coefficient matrix \begin{align*}
L\ \left({\boldsymbol{D,W,}{\boldsymbol{V}_4}} \right) & = \left\| {\boldsymbol{Y} - \boldsymbol{PDA}} \right\|_F^2 + \left\| {\boldsymbol{X} - \boldsymbol{DAH}} \right\|_F^2 \\
&\quad + \mu \left\| {\boldsymbol{W} - \boldsymbol{D} + \frac{{{\boldsymbol{V}_4}}}{{2\mu }}} \right\|_F^2\ \\
& {\rm{s}}{\rm{.t}}.\,0 \leq \boldsymbol{W} \leq 1 \tag{31}
\end{align*}
\begin{align*}
{\boldsymbol{D}^{\left({\boldsymbol{t} + 1} \right)}} = & {\rm{arg}}\mathop {\min }\limits_{\boldsymbol D} L\left({\boldsymbol{D,}{\boldsymbol{W}^{\left(\boldsymbol{t} \right)}},\boldsymbol{V}_4^{\left(\boldsymbol{t} \right)}} \right)\\
{\boldsymbol{W}^{\left({\boldsymbol{t} + 1} \right)}} = & {\rm{arg}}\mathop {\min }\limits_{\boldsymbol W} L\left({{\boldsymbol{D}^{\left(\boldsymbol{t} \right)}},\boldsymbol{W},\boldsymbol{V}_4^{\left(\boldsymbol{t} \right)}} \right). \tag{32}
\end{align*}
Meanwhile, the Lagrangian multiplier is updated by
\begin{equation*}
\boldsymbol{V}_4^{\left({\boldsymbol{t} + 1} \right)} = \boldsymbol{V}_4^{\left(\boldsymbol{t} \right)}\ + \mu \left({{\boldsymbol{W}^{\left({\boldsymbol{t} + 1} \right)}} - {\boldsymbol{D}^{\left({\boldsymbol{t} + 1} \right)}}} \right). \tag{33}
\end{equation*}
The two subproblems in (32) can be easily solved analytically. For the updating of \begin{equation*}
{\boldsymbol{D}^{\left({\boldsymbol{t} + 1} \right)}}{\boldsymbol{H}_1} + {\boldsymbol{H}_2}{\boldsymbol{D}^{\left({\boldsymbol{t} + 1} \right)}} = {\boldsymbol{H}_3}\tag{34}
\end{equation*}
\begin{equation*}
\begin{array}{l} {\boldsymbol{H}_1} = \left[ {\left({\boldsymbol{AH}} \right){{\left({\boldsymbol{AH}} \right)}^T} + \mu \boldsymbol{I}} \right]\ {\left({\boldsymbol{A}{\boldsymbol{A}^T}} \right)^{ - 1}}\\ {\boldsymbol{H}_2} = {\boldsymbol{P}^T}\boldsymbol{P}\\ {\boldsymbol{H}_3} = \left[ {\boldsymbol{X}{{\left({\boldsymbol{AH}} \right)}^T} + {\boldsymbol{P}^T}\boldsymbol{Y}{\boldsymbol{A}^T} + \mu \left({{\boldsymbol{W}^{\left(\boldsymbol{t} \right)}} + \frac{{\boldsymbol{V}_4^{\left(\boldsymbol{t} \right)}}}{{2\mu }}} \right)} \right]\ {\left({\boldsymbol{A}{\boldsymbol{A}^T}} \right)^{ - 1}}. \end{array} \tag{35}
\end{equation*}
Then, vectorizing \begin{equation*}
\left({\boldsymbol{H}_1^T \otimes \boldsymbol{I} + I \otimes {\boldsymbol{H}_2}} \right){\rm{vec\ }}\left({{\boldsymbol{D}^{\left({\boldsymbol{t} + 1} \right)}}} \right) = \ {\rm{vec}}\left({{\boldsymbol{H}_3}} \right) \tag{36}
\end{equation*}
\begin{equation*}
{\rm{vec\ }}\left({{\boldsymbol{D}^{\left({\boldsymbol{t} + 1} \right)}}} \right) = {\left({\boldsymbol{H}_1^T \otimes \boldsymbol{I} + \boldsymbol{I} \otimes {\boldsymbol{H}_2}} \right)^{ - 1}}{\rm{\ vec}}\left({{\boldsymbol{H}_3}} \right). \tag{37}
\end{equation*}
For the updating of \begin{equation*}
{\boldsymbol{W}^{\left({\boldsymbol{t} + 1} \right)}} = \ {\rm{min}}\left({{\rm{max}}\left({{\boldsymbol{D}^{\left({\boldsymbol{t} + 1} \right)}} - \frac{{\boldsymbol{V}_4^{\left(\boldsymbol{t} \right)}}}{{2\mu }},0} \right),1} \right). \tag{38}
\end{equation*}
The update of variables and multiplier is alternately iterated until convergence. The overall algorithm for updating spectral basis
Algorithm 3: Update \boldsymbol{D} With \boldsymbol{A} Fixed.
Input: LR-HSI
Initialization:
While not converge do
Update variables
Update Lagrangian multipliers
Update
End while
Output: spectral basis
Experimental Results and Discussion
In this section, to evaluate the performance of our proposed HSI super-resolution method, we conduct ample experiments on both ground-based HSIs datasets and real remote sensing HSI. To objectively evaluate the quality of the reconstructed HSIs, we adopt four objective evaluation indices, which are peak signal to noise ratio (PSNR), root-mean-square error (RMSE), relative dimensionless global error in synthesis (ERGAS), and spectral angle mapper (SAM), in our experiments.
A. Experimental Datasets
In our experiments, we use two categories of images to show the effectiveness of our method. For the ground-based HSIs, we use a public HSI dataset, which is named Columbia Computer Vision Laboratory (CAVE) [65]. The CAVE dataset includes 32 HSIs of everyday objects, which are captured by generalized assorted pixel camera with high quality. The spatial size of each HSI in CAVE is
Total of 20 representative testing images from the CAVE datasets. (a) Oil_painting. (b) Cloth. (c) Fake_and_real_peppers. (d) Balloons. (e) Fake_and_real_food. (f) Beads. (g) CD. (h) Chart_and_stuffed_toy. (i) Egyptain_statue. (j) Face. (k) Fake_and_real_lemon_slices. (l) Fake_and_real_sushi. (m) Feathers. (n) Flowers. (o) Glass_tiles. (p) Paints. (q) Real_and_fake_apples. (r) Sponges. (s) Stuffed_toys. (t) Thread_spools.
For the real remote sensing HSI, we use three popular remote sensing HSIs: Cuprite Mine Nevada, Indian Pines, and Pavia Center, which are adopted in [51]. The three HSIs are shown in Fig. 5. The wavelength of the Cuprite Mine Nevada image ranges from 400 to 2500 nm at an interval of 10 nm and the spatial resolution of the Cuprite mine Nevada image is 20 m. We crop the top left region of size
Three popular remote sensing HSIs. (a) Cuprite Mine Nevada. (b) Indian Pines. (c) Pavia Center.
B. Evaluation Indices
In this article, we use four indices to evaluate the reconstruction quality. The first index is PSNR, which is defined as the average PSNR value of all spectral bands. The formulation is as follows:
\begin{equation*}
{\rm{PSNR}}\left({\hat{\boldsymbol{Z}},\boldsymbol{Z}} \right) = \frac{1}{S}\mathop \sum \limits_{i\ = {\rm{\ }}1}^S {\rm{PSNR}}\left({{{\hat{\boldsymbol{Z}}}_{\boldsymbol i}}\boldsymbol{,}{\boldsymbol{Z}_{\boldsymbol i}}} \right) \tag{39}
\end{equation*}
The second index is RMSE, which is defined as the average RMSE of all spectral bands, i.e.,
\begin{equation*}
{\rm{RMSE\ }}\left({\hat{\boldsymbol{Z},Z}} \right) = \frac{1}{S}\ \mathop \sum \limits_{i\ = {\rm{\ }}1}^S {\rm{RMSE}}\left({{{\hat{\boldsymbol{Z}}}_{\boldsymbol i}}\boldsymbol{,}{\boldsymbol{Z}_{\boldsymbol i}}} \right). \tag{40}
\end{equation*}
The smaller the RMSE, the better the reconstruction result.
The third index is ERGAS, whose formulation is
\begin{equation*}
{\rm{ERGAS\ }}\left({\hat{\boldsymbol{Z},Z}} \right) = \frac{{100}}{c}\ \sqrt {\frac{1}{S}\mathop \sum \nolimits_{i\ = {\rm{\ }}1}^S \frac{{{\rm{MSE}}\left({{{\hat{\boldsymbol{Z}}}_{\boldsymbol i}}\boldsymbol{,}{\boldsymbol{Z}_{\boldsymbol i}}} \right)}}{{\mu _{{{\hat{\boldsymbol{Z}}}_{\boldsymbol i}}}^2}}} \tag{41}
\end{equation*}
The fourth index is SAM, which is defined as
\begin{equation*}
{\rm{SAM\ }}\left({\hat{\boldsymbol Z}, {\boldsymbol Z}} \right) = \frac{1}{N}\ \mathop {\sum }_{j\ = \ 1}^N {\cos ^{-1}} \frac {{\hat{\boldsymbol z}_{\boldsymbol j}^{\boldsymbol T}{{\boldsymbol z}_{\boldsymbol j}}}}{{{{\left\| {{\hat{\boldsymbol z}}_{\boldsymbol j}} \right\|}_2}{{\left\| {{{\boldsymbol z}_{\boldsymbol j}}} \right\|}_2}}} \tag{42}
\end{equation*}
C. Experimental Settings for the Comparison Methods
For the sake of fairness, we describe the experimental settings in this section. The LR-HSI
For the ground-based HSIs, as in [52], the ground truth HR-HSI
For the real remote sensing HSIs, as the operations in [51], the ground truth HR-HSI
D. Experimental Results of Our Method
In this section, we will show the experimental results of our HSI super-resolution method compared with some typical existing HSI super-resolution methods, including G-SOMP+ [49], SNNMF [66], CNMF [43], NSSR [52], SSR [51], and OTD [48]. Note that we do not compare our method with SSR in the experiments of the ground-based HSIs. We will explain it in the experimental analysis of real remote sensing HSI. Similarly, we do not compare our method with OTD in the experiments of the ground-based HSIs. It is because that Han et al. [48] only conducts experiments with real remote sensing HSIs. For the sake of fairness, we only compared our method with OTD on the real remote sensing HSIs. In recent years, deep learning based HSI super-resolution methods have shown good results, but this kind of method needs a large number of training samples. Considering that our method do not need any training sample, we do not compare our approach with the deep learning based methods. In our experiments, we assume that the spatial degradation operator
For the ground-based HSIs, the comparison of results is given in Table I. The best results are bolded for clarity. As presented in Table I, our ANSR method achieves the best results among all the compared methods, and the NSSR is the second-best method, although it was proposed in 2016. According to Table I, the large average PSNR gains of our method over the second-best method with
PSNR curves of all the wavelengths of spectral bands over the testing image “fake_and_real_food.” (a) PNSR curves with scaling factor
Reconstructed results of image “fake_and_real_peppers” at 480, 550, and 640 nm with scaling factor
Reconstructed results of image “Egyptian_statue” at 480, 550, and 640 nm with scaling factor
Reconstructed results of image “real_and_fake_apples” at 480, 550, and 640 nm with scaling factor
For the real remote sensing HSI, the comparison of results is given in Tables II–IV. The best results are bolded for clarity. In this part, we compare our ANSR method with SSR and OTD, which are not compared in the experiments of the ground-based HSIs. It is because that the SSR method needs to cluster the HR-MSI into superpixels, whose size and shape are adaptively adjusted according to the local structures. However, the images in the CAVE dataset are always very simple, and there is not much information in them. Therefore, the SSR method will fall into an endless loop because some superpixels only contain invalid information, when the number of superpixels is too large (e.g., 6000, which is used in real remote sensing HSI). But if we reduce the number of superpixels, the results of the SSR method become very poor because it loses its advantage. As presented in Tables II–IV, our ANSR method performs best among all the compared methods and the OTD is the second-best method. According to Table II, the large PSNR gains of our method over the second-best method on the “Cuprite Mine Nevada” with
Reconstructed results of image “Pavia Center” at
According to the above experimental discussion, we can find that the larger the scaling factor, the more obvious the advantage of our method. In our experiments, the PSNR values of our method have the most obvious advantage over the second-best method when scaling factor
E. Parameters Selection in Our Method
In our HSI super-resolution method, there are three crucial parameters, i.e., regularization parameters
First, we all know that the number of atoms of the basis is very important in a sparse coding-based method. We perform some experiments with different values of
PSNR curves of the ground-based HSIs, Cuprite Mine Nevada, Indian Pines, and Pavia Center as functions of the number of atoms
Second, we test and verify the effect of
PSNRs of the ground-based HSIs, Cuprite Mine Nevada, Indian Pines, and Pavia Center as functions of
Third, we test and verify the effect of
PSNRs of the ground-based HSIs, Cuprite Mine Nevada, Indian Pines, and Pavia Center as functions of
Conclusion
In this article, we presented a novel sparse representation-based HSI super-resolution method, termed ANSR, to fuse an LR-HSI and its corresponding HR-MSI. In the base of the NSSR model, we introduce the ASR, which can balance the relationship between the sparsity and collaboration by generating a suitable coefficient, into our ASNR method. Also, we design an alternative optimization algorithm to optimize the spectral basis rather than keeping it fixed. ADMM method is applied to solve the proposed optimization problem. In order to show the performance of the proposed method, we conduct plenty of experiments. The experimental results on both ground-based HSI dataset and real remote sensing HSIs show the superiority of our proposed approach to some other state-of-the-art HSI super-resolution methods.
In further work, we aim to improve the method in several directions. We will focus on the estimation of the spatial degradation operator