I. Introduction
Image super-resolution (SR) aims to restore high-resolution (HR) images from given low-resolution (LR) images. Over the past few years, numerous convolutional neural network (CNN)-based SR methods have been proposed [1], [2], [3] and achieved remarkable performance. However, usually, the SR methods are applied to the standard RGB (sRGB) images produced by the image signal processing (ISP) pipeline of digital cameras [4], thus leading to the following three main drawbacks:
Error accumulation: Image restoration operations, including color demosaicking [5], [6], [7], [8], [9] and noise reduction [10], [11], [12] (and sometimes deblurring), have been applied in the ISP pipeline [13]. Moreover, sRGB images are usually compressed and stored, e.g., in JPEG format, and thus inevitably brings quantization noises and compression artifacts. Since these operations are processed separately, the errors produced by each operation may accumulate and spread within sRGB images, potentially affecting the SR method.
Low bit depth: Typically, in sRGB images, only 8 bits are used to record one color channel of a single pixel. Compared to a higher bit depth, such as 12 or 14 bits, the relatively low bit depth used in sRGB images limits the recorded visual information and thus considerably restricts the quality of SR results.
Nonlinearity with scene radiance: In a common ISP pipeline, nonlinear operations, such as tone curve, look-up table (LUT) and gamma transformation, are applied to convert the colors from the linear color space to the nonlinear sRGB space [14], [15]. As a result, the pixel intensity in sRGB images is no longer linear to the scene radiance [16], which makes the SR task even more challenging.