PRNet: Pyramid Restoration Network for RAW Image Super-Resolution | IEEE Journals & Magazine | IEEE Xplore

PRNet: Pyramid Restoration Network for RAW Image Super-Resolution


Abstract:

Typically, image super-resolution (SR) methods are applied to the standard RGB (sRGB) images produced by the image signal processing (ISP) pipeline of digital cameras. Ho...Show More

Abstract:

Typically, image super-resolution (SR) methods are applied to the standard RGB (sRGB) images produced by the image signal processing (ISP) pipeline of digital cameras. However, due to error accumulation, low bit depth and the nonlinearity with scene radiance in sRGB images, performing SR on them is sub-optimal. To address this issue, a RAW image SR method called pyramid restoration network (PRNet) is proposed in this paper. Firstly, PRNet takes the low-resolution (LR) RAW image as input, and generates a rough estimation of the SR result in the linear color space. Afterwards, a pyramid refinement (PR) sub-network refines image details in the intermediate SR result and corrects its colors in a divide-and-conquer manner. To learn the appropriate colors for displaying, external guidance is extracted from the LR reference image in the sRGB color space, and then fed to the PR sub-network. To effectively incorporate the external guidance, the cross-layer correction module (CLCM), which fully investigates the long-range interactions between two input features, is introduced in the PR sub-network. Moreover, as different frequency components decomposed from the same image are highly correlated, in the PR sub-network, the refined features from a lower layer are utilized to support the feature refinement in an upper layer. Extensive experiments presented in this paper demonstrate that the proposed method is capable of recovering fine details and small structures in images while producing vivid colors that align with the output of a specific camera ISP pipeline.
Published in: IEEE Transactions on Computational Imaging ( Volume: 10)
Page(s): 479 - 495
Date of Publication: 12 March 2024

ISSN Information:

Funding Agency:

References is not available for this document.

I. Introduction

Image super-resolution (SR) aims to restore high-resolution (HR) images from given low-resolution (LR) images. Over the past few years, numerous convolutional neural network (CNN)-based SR methods have been proposed [1], [2], [3] and achieved remarkable performance. However, usually, the SR methods are applied to the standard RGB (sRGB) images produced by the image signal processing (ISP) pipeline of digital cameras [4], thus leading to the following three main drawbacks:

Error accumulation: Image restoration operations, including color demosaicking [5], [6], [7], [8], [9] and noise reduction [10], [11], [12] (and sometimes deblurring), have been applied in the ISP pipeline [13]. Moreover, sRGB images are usually compressed and stored, e.g., in JPEG format, and thus inevitably brings quantization noises and compression artifacts. Since these operations are processed separately, the errors produced by each operation may accumulate and spread within sRGB images, potentially affecting the SR method.

Low bit depth: Typically, in sRGB images, only 8 bits are used to record one color channel of a single pixel. Compared to a higher bit depth, such as 12 or 14 bits, the relatively low bit depth used in sRGB images limits the recorded visual information and thus considerably restricts the quality of SR results.

Nonlinearity with scene radiance: In a common ISP pipeline, nonlinear operations, such as tone curve, look-up table (LUT) and gamma transformation, are applied to convert the colors from the linear color space to the nonlinear sRGB space [14], [15]. As a result, the pixel intensity in sRGB images is no longer linear to the scene radiance [16], which makes the SR task even more challenging.

Select All
1.
C. Dong, C. C. Loy, K. He and X. Tang, "Image super-resolution using deep convolutional networks", IEEE Trans. Pattern Anal. Mach. Intell., vol. 38, no. 2, pp. 295-307, Feb. 2016.
2.
C. Dong, C. C. Loy and X. Tang, "Accelerating the super-resolution convolutional neural network", Proc. Eur. Conf. Comput. Vis., pp. 391-407, 2016.
3.
Z. Wang, J. Chen and S. C. H. Hoi, "Deep learning for image super-resolution: A survey", IEEE Trans. Pattern Anal. Mach. Intell., vol. 43, no. 10, pp. 3365-3387, Oct. 2021.
4.
S. W. Hasinoff et al., "Burst photography for high dynamic range and low-light imaging on mobile cameras", ACM Trans. Graph., vol. 35, no. 6, pp. 1-12, Dec. 2016.
5.
M. Gharbi, G. Chaurasia, S. Paris and F. Durand, "Deep joint demosaicking and denoising", ACM Trans. Graph., vol. 35, no. 6, pp. 1-12, Dec. 2016.
6.
Y. Tan, K. Chang, H. Li, Z. Tang and T. Qin, "Lightweight color image demosaicking with multi-core feature extraction", Proc. IEEE Conf. Vis. Commun. Image Process., pp. 136-139, 2020.
7.
L. Liu, X. Jia, J. Liu and Q. Tian, "Joint demosaicing and denoising with self guidance", Proc. IEEE Conf. Comput. Vis. Pattern Recognit., pp. 2237-2246, 2020.
8.
T. Zhang, Y. Fu and C. Li, "Deep spatial adaptive network for real image demosaicing", Proc. AAAI Conf. Artif. Intell., pp. 3326-3334, 2022.
9.
K. Feng et al., "Mosaic convolution-attention network for demosaicing multispectral filter array images", IEEE Trans. Comput. Imag., vol. 7, pp. 864-878, 2021.
10.
T. Brooks et al., "Unprocessing images for learned raw denoising", Proc. IEEE Conf. Comput. Vis. Pattern Recognit., pp. 11028-11037, 2019.
11.
J. Liu et al., "Learning raw image denoising with Bayer pattern unification and Bayer preserving augmentation", Proc. IEEE Conf. Comput. Vis. Pattern Recognit. Workshops, pp. 2070-2077, 2019.
12.
O. A. Elgendy, A. Gnanasambandam, S. H. Chan and J. Ma, "Low-light demosaicking and denoising for small pixels using learned frequency selection", IEEE Trans. Comput. Imag., vol. 7, pp. 137-150, 2021.
13.
K. Chang, H. Li, Y. Tan, P. L. K. Ding and B. Li, "A two-stage convolutional neural network for joint demosaicking and super-resolution", IEEE Trans. Circuits Syst. Video Technol., vol. 32, no. 7, pp. 4238-4254, Jul. 2022.
14.
H. C. Karaimer and M. S. Brown, "A software platform for manipulating the camera imaging pipeline", Proc. Eur. Conf. Comput. Vis., pp. 429-444, 2016.
15.
Y. Tang, K. Chang, M. Huang and B. Li, "BMISP: Bidirectional mapping of image signal processing pipeline", Signal Process., vol. 212, Nov. 2023.
16.
R. M. H. Nguyen and M. S. Brown, "RAW image reconstruction using a self-contained sRGB-JPEG image with only 64 KB overhead", Proc. IEEE Conf. Comput. Vis. Pattern Recognit., pp. 1655-1663, 2016.
17.
R. Zhou, R. Achanta and S. Süsstrunk, "Deep residual network for joint demosaicing and super-resolution", Proc. Color Imag. Conf., pp. 75-80, 2018.
18.
X. Xu, Y. Ye and X. Li, "Joint demosaicing and super-resolution (JDSR): Network design and perceptual optimization", IEEE Trans. Comput. Imag., vol. 6, pp. 968-980, 2020.
19.
W. Xing and K. Egiazarian, "End-to-end learning for joint image demosaicing denoising and super-resolution", Proc. IEEE Conf. Comput. Vis. Pattern Recognit., pp. 3506-3515, 2021.
20.
G. Qian et al., "Rethinking learning-based demosaicing denoising and super-resolution pipeline", Proc. IEEE Int. Conf. Comput. Photogr., pp. 1-12, 2022.
21.
X. Xu, Y. Ma, W. Sun and M.-H. Yang, "Exploiting raw images for real-scene super-resolution", IEEE Trans. Pattern Anal. Mach. Intell., vol. 44, no. 4, pp. 1905-1921, Apr. 2022.
22.
P. J. Burt and E. H. Adelson, "The Laplacian pyramid as a compact image code", IEEE Trans. Commun., vol. 31, no. 4, pp. 532-540, Apr. 1983.
23.
X. Wang, K. Yu, C. Dong and C. Change Loy, "Recovering realistic texture in image super-resolution by deep spatial feature transform", Proc. IEEE Conf. Comput. Vis. Pattern Recognit., pp. 606-615, 2018.
24.
K. He, X. Zhang, S. Ren and J. Sun, "Deep residual learning for image recognition", Proc. IEEE Conf. Comput. Vis. Pattern Recognit., pp. 770-778, 2016.
25.
B. Lim, S. Son, H. Kim, S. Nah and K. M. Lee, "Enhanced deep residual networks for single image super-resolution", Proc. IEEE Conf. Comput. Vis. Pattern Recognit. Workshops, pp. 1132-1140, 2017.
26.
X. Wang et al., "ESRGAN: Enhanced super-resolution generative adversarial networks", Proc. Eur. Conf. Comput. Vis. Workshops, pp. 63-79, 2018.
27.
N. Ahn, B. Kang and K.-A. Sohn, "Fast accurate and lightweight super-resolution with cascading residual network", Proc. Eur. Conf. Comput. Vis., pp. 256-272, 2018.
28.
Y. Tai, J. Yang, X. Liu and C. Xu, "MemNet: A persistent memory network for image restoration", Proc. IEEE Int. Conf. Comput. Vis., pp. 4549-4557, 2017.
29.
T. Tong, G. Li, X. Liu and Q. Gao, "Image super-resolution using dense skip connections", Proc. IEEE Int. Conf. Comput. Vis., pp. 4809-4817, 2017.
30.
Y. Zhang, Y. Tian, Y. Kong, B. Zhong and Y. Fu, "Residual dense network for image super-resolution", Proc. IEEE Conf. Comput. Vis. Pattern Recognit., pp. 2472-2481, 2018.
Contact IEEE to Subscribe

References

References is not available for this document.