Loading [MathJax]/extensions/MathZoom.js
UNFOLD: 3-D U-Net, 3-D CNN, and 3-D Transformer-Based Hyperspectral Image Denoising | IEEE Journals & Magazine | IEEE Xplore

UNFOLD: 3-D U-Net, 3-D CNN, and 3-D Transformer-Based Hyperspectral Image Denoising


Abstract:

Hyperspectral images (HSIs) encompass data across numerous spectral bands, making them valuable in various practical fields such as remote sensing, agriculture, and marin...Show More

Abstract:

Hyperspectral images (HSIs) encompass data across numerous spectral bands, making them valuable in various practical fields such as remote sensing, agriculture, and marine monitoring. Unfortunately, inevitable noise introduction during sensing restricts their applicability, necessitating denoising for optimal utilization. The existing deep learning (DL)-based denoising methods suffer from various limitations. For instance, convolutional neural networks (CNNs) struggle with long-range dependencies, while vision transformers (ViTs) struggle to capture local details. This article introduces a novel method, UNFOLD, that addresses these inherent limitations by harmoniously integrating the strengths of 3-D U-Net, 3-D CNN, and 3-D Transformer architectures. Unlike several existing methods that predominantly capture dependencies either along the spatial or the spectral dimension, UNFOLD addresses HSI denoising as a 3-D task, synergizing spatial and spectral information through the utilization of 3-D Transformer and 3-D CNN. It employs the self-attention (SA) mechanism of Transformers to capture the global dependencies and model long-range relationships across spatial and spectral dimensions. To overcome the limitations of 3-D Transformer in capturing fine-grained local and spatial features, UNFOLD complements it by incorporating 3-D CNN. Moreover, UNFOLD utilize a modified form of 3-D U-Net architecture for HSI denoising, wherein it employs a 3-D Transformer-based encoder instead of the conventional 3-D CNN-based encoder. It further capitalizes on the property of U-Net to integrate features across various scales, thereby enhancing efficacy by preserving intricate structural details. Results from extensive experiments demonstrate that UNFOLD outperforms the state-of-the-art HSI denoising methods.
Article Sequence Number: 5529710
Date of Publication: 31 October 2023

ISSN Information:

Funding Agency:

References is not available for this document.

I. Introduction

Hyperspectral image (HSI) contains information at several spectrums, thus, extensively used in several real-world domains, including remote sensing [1], classification [2], [3], [4], agriculture [5], and marine monitoring [6]. It is represented as a 3-D array, incorporating two spatial and one spectral dimension. Unfortunately, noise can be added during the HSI sensing due to various factors, including limited light, photon effects, and atmospheric interference [7], thereby degrading HSI quality. This issue is mitigated by HSI denoising. In computer vision, image denoising is performed by analyzing each pixel’s behavior with respect to its local neighborhood or global context. Several linear filters are extensively employed in the literature for local neighborhood analysis. Similarly, the nonlocal means filter [8] and non-local meets Global (NGMeet) [9] have been utilized to analyze global context or long-range dependencies for denoising [10]. Hence, image denoising can be effectively performed by analyzing the local neighborhood and global context.

Select All
1.
F. Melgani and L. Bruzzone, "Classification of hyperspectral remote sensing images with support vector machines", IEEE Trans. Geosci. Remote Sens., vol. 42, no. 8, pp. 1778-1790, Aug. 2004.
2.
H. Zhang, Y. Li, Y. Jiang, P. Wang, Q. Shen and C. Shen, "Hyperspectral classification based on lightweight 3-D-CNN with transfer learning", IEEE Trans. Geosci. Remote Sens., vol. 57, no. 8, pp. 5813-5828, Aug. 2019.
3.
Y. Zhang, W. Li, W. Sun, R. Tao and Q. Du, "Single-source domain expansion network for cross-scene hyperspectral image classification", IEEE Trans. Image Process., vol. 32, pp. 1498-1512, 2023.
4.
M. Zhang, X. Zhao, W. Li, Y. Zhang, R. Tao and Q. Du, "Cross-scene joint classification of multisource data with multilevel domain adaption network", IEEE Trans. Neural Netw. Learn. Syst., Apr. 2023.
5.
K. Bi, S. Xiao, S. Gao, C. Zhang, N. Huang and Z. Niu, "Estimating vertical chlorophyll concentrations in maize in different health states using hyperspectral LiDAR", IEEE Trans. Geosci. Remote Sens., vol. 58, no. 11, pp. 8125-8133, Nov. 2020.
6.
Z. Ping Lee, W. J. Rhea, R. Arnone and W. Goode, "Absorption coefficients of marine waters: Expanding multiband information to hyperspectral data", IEEE Trans. Geosci. Remote Sens., vol. 43, no. 1, pp. 118-124, Jan. 2005.
7.
Q. Yuan, Q. Zhang, J. Li, H. Shen and L. Zhang, "Hyperspectral image denoising employing a spatial–spectral deep residual convolutional neural network", IEEE Trans. Geosci. Remote Sens., vol. 57, no. 2, pp. 1205-1218, Feb. 2019.
8.
A. Buades, B. Coll and J. M. Morel, "A review of image denoising algorithms with a new one", Multiscale Model. Simul., vol. 4, no. 2, pp. 490-530, Jan. 2005.
9.
W. He, Q. Yao, C. Li, N. Yokoya and Q. Zhao, "Non-local meets global: An integrated paradigm for hyperspectral denoising", Proc. IEEE/CVF Conf. Comput. Vis. Pattern Recognit. (CVPR), pp. 6861-6870, Jun. 2019.
10.
Y. Qian, Y. Shen, M. Ye and Q. Wang, "3-D nonlocal means filter with noise estimation for hyperspectral imagery denoising", Proc. IEEE Int. Geosci. Remote Sens. Symp., pp. 1345-1348, Jul. 2012.
11.
K. Dabov, A. Foi, V. Katkovnik and K. Egiazarian, "Image denoising by sparse 3-D transform-domain collaborative filtering", IEEE Trans. Image Process., vol. 16, no. 8, pp. 2080-2095, Aug. 2007.
12.
H. Othman and S.-E. Qian, "Noise reduction of hyperspectral imagery using hybrid spatial–spectral derivative-domain wavelet shrinkage", IEEE Trans. Geosci. Remote Sens., vol. 44, no. 2, pp. 397-408, Feb. 2006.
13.
Y. Wang, J. Peng, Q. Zhao, Y. Leung, X.-L. Zhao and D. Meng, "Hyperspectral image restoration via total variation regularized low-rank tensor decomposition", IEEE J. Sel. Topics Appl. Earth Observ. Remote Sens., vol. 11, no. 4, pp. 1227-1243, Apr. 2018.
14.
H. Chen, G. Yang and H. Zhang, "Hider: A hyperspectral image denoising transformer with spatial–spectral constraints for hybrid noise removal", IEEE Trans. Neural Netw. Learn. Syst., Oct. 2022.
15.
X. Cao, X. Fu, C. Xu and D. Meng, "Deep spatial–spectral global reasoning network for hyperspectral image denoising", IEEE Trans. Geosci. Remote Sens., vol. 60, 2022.
16.
K. Wei, Y. Fu and H. Huang, "3-D quasi-recurrent neural network for hyperspectral image denoising", IEEE Trans. Neural Netw. Learn. Syst., vol. 32, no. 1, pp. 363-375, Jan. 2021.
17.
O. Ronneberger, P. Fischer and T. Brox, "U-Net: Convolutional networks for biomedical image segmentation", Proc. Int. Conf. Med. Image Comput. Comput.-Assist. Intervent., pp. 234-241, 2015.
18.
H. Zeng et al., "Degradation-noise-aware deep unfolding transformer for hyperspectral image denoising", arXiv:2305.04047, 2023.
19.
A. Dosovitskiy et al., "An image is worth 16×16 words: Transformers for image recognition at scale", arXiv:2010.11929, 2020.
20.
A. Vaswani et al., "Attention is all you need", Proc. Adv. Neural Inf. Process. Syst., vol. 30, pp. 5998-6008, 2017.
21.
T. Wang et al., "O-Net: A novel framework with deep fusion of CNN and transformer for simultaneous segmentation and classification", Frontiers Neurosci., vol. 16, Jun. 2022.
22.
Y. Chen, Y. Guo, Y. Wang, D. Wang, C. Peng and G. He, "Denoising of hyperspectral images using nonconvex low rank matrix approximation", IEEE Trans. Geosci. Remote Sens., vol. 55, no. 9, pp. 5366-5380, Sep. 2017.
23.
L. Zhuang and J. M. Bioucas-Dias, "Fast hyperspectral image denoising and inpainting based on low-rank and sparse representations", IEEE J. Sel. Topics Appl. Earth Observ. Remote Sens., vol. 11, no. 3, pp. 730-742, Mar. 2018.
24.
B. Du, Z. Huang, N. Wang, Y. Zhang and X. Jia, "Joint weighted nuclear norm and total variation regularization for hyperspectral image denoising", Int. J. Remote Sens., vol. 39, no. 2, pp. 334-355, Jan. 2018.
25.
Y. Chang, L. Yan and S. Zhong, "Hyper-Laplacian regularized unidirectional low-rank tensor recovery for multispectral image denoising", Proc. IEEE Conf. Comput. Vis. Pattern Recognit. (CVPR), pp. 5901-5909, Jul. 2017.
26.
A. Voulodimos, N. Doulamis, A. Doulamis and E. Protopapadakis, "Deep learning for computer vision: A brief review", Comput. Intell. Neurosci., vol. 2018, pp. 1-13, Jan. 2018.
27.
T. Bodrito, A. Zouaoui, J. Chanussot and J. Mairal, "A trainable spectral–spatial sparse coding model for hyperspectral image restoration", Proc. Adv. Neural Inf. Process. Syst., vol. 34, pp. 5430-5442, 2021.
28.
A. K. Gupta, R. Kumar, L. Birla and P. Gupta, "RADIANT: Better rPPG estimation using signal embeddings and transformer", Proc. IEEE/CVF Winter Conf. Appl. Comput. Vis. (WACV), pp. 4965-4975, Jan. 2023.
29.
M. Wang, W. He and H. Zhang, "A spatial–spectral transformer network with total variation loss for hyperspectral image denoising", IEEE Geosci. Remote Sens. Lett., vol. 20, pp. 1-5, 2023.
30.
M. Li, Y. Fu and Y. Zhang, "Spatial–spectral transformer for hyperspectral image denoising", Proc. AAAI Conf. Artif. Intell., vol. 37, no. 1, pp. 1368-1376, 2023.

Contact IEEE to Subscribe

References

References is not available for this document.