Loading [MathJax]/extensions/MathZoom.js
Fine-Grained Attention and Feature-Sharing Generative Adversarial Networks for Single Image Super-Resolution | IEEE Journals & Magazine | IEEE Xplore

Fine-Grained Attention and Feature-Sharing Generative Adversarial Networks for Single Image Super-Resolution


Abstract:

Traditional super-resolution (SR) methods by minimize the mean square error usually produce images with over-smoothed and blurry edges, due to the lack of high-frequency ...Show More

Abstract:

Traditional super-resolution (SR) methods by minimize the mean square error usually produce images with over-smoothed and blurry edges, due to the lack of high-frequency details. In this paper, we propose two novel techniques within the generative adversarial network framework to encourage generation of photo-realistic images for image super-resolution. Firstly, instead of producing a single score to discriminate real and fake images, we propose a variant, called Fine-grained Attention Generative Adversarial Network (FASRGAN), to discriminate each pixel of real and fake images. FASRGAN adopts a UNet-like network as the discriminator with two outputs: an image score and an image score map. The score map has the same spatial size as the HR/SR images, serving as the fine-grained attention to represent the degree of reconstruction difficulty for each pixel. Secondly, instead of using different networks for the generator and the discriminator, we introduce a feature-sharing variant (denoted as Fs-SRGAN) for both the generator and the discriminator. The sharing mechanism can maintain model express power while making the model more compact, and thus can improve the ability of producing high-quality images. Quantitative and visual comparisons with state-of-the-art methods on benchmark datasets demonstrate the superiority of our methods. We further apply our super-resolution images for object recognition, which further demonstrates the effectiveness of our proposed method. The code is available at https://github.com/Rainyfish/FASRGAN-and-Fs-SRGAN.
Published in: IEEE Transactions on Multimedia ( Volume: 24)
Page(s): 1473 - 1487
Date of Publication: 12 March 2021

ISSN Information:

Funding Agency:

References is not available for this document.

I. Introduction

Single image super-resolution (SISR), which aims to recover a high-resolution (HR) image from its low-solution (LR) version, has been an active research topic in computer graphic and vision for decades. SISR has also attracted increasing attention in both academia and industry, with applications in various fields such as medical imaging, security surveillance, object recognition and so on. However, SISR is a typically ill-posed problem due to the irreversible image degradation process, i.e., multiple HR images can be generated from one single LR image. Learning the mapping between HR and LR images plays an important role in addressing this problem.

Select All
1.
C. Dong, C. C. Loy, K. He and X. Tang, "Image super-resolution using deep convolutional networks" in IEEE Trans. Pattern Anal. Mach. Intell., vol. 38, no. 2, pp. 295-307, Feb. 2015.
2.
Z. Wang, J. Chen and S. C. H. Hoi, "Deep learning for image super-resolution: A survey", IEEE Trans. Pattern Anal. Mach. Intell., pp. 1-22.
3.
N. Kumar and A. Sethi, "Fast learning-based single image super-resolution", IEEE Trans. Multimedia, vol. 18, no. 8, pp. 1504-1515, Aug. 2016.
4.
J. Kim, J. K. Lee and K. M. Lee, "Accurate image super-resolution using very deep convolutional networks", Proc. IEEE Conf. Comput. Vis. Pattern Recognit., pp. 1646-1654, Jun. 2016.
5.
B. Lim, S. Son, H. Kim, S. Nah and K. Mu Lee, "Enhanced deep residual networks for single image super-resolution", Proc. IEEE Conf. Comput. Vis. Pattern Recognit. Workshops, pp. 136-144, 2017.
6.
K. He, X. Zhang, S. Ren and J. Sun, "Deep residual learning for image recognition", Proc. IEEE Conf. Comput. Vis. Pattern Recognit., pp. 770-778, 2016.
7.
Y. Zhang et al., "Image super-resolution using very deep residual channel attention networks", Proc. Eur. Conf. Comput. Vis., pp. 286-301, 2018.
8.
C. Ledig et al., "Photo-realistic single image super-resolution using a generative adversarial network", Proc. IEEE Conf. Comput. Vis. Pattern Recognit., pp. 4681-4690, 2017.
9.
I. Goodfellow et al., "Generative adversarial nets", Proc. Adv. Neural Inf. Process. Syst., pp. 2672-2680, 2014.
10.
X. Wang et al., "ESRGAN: Enhanced super-resolution generative adversarial networks", Proc. Eur. Conf. Comput. Vis., pp. 63-79, 2018.
11.
A. Jolicoeur-Martineau, "The relativistic discriminator: A key element missing from standard GAN", Proc. Int. Conf. Learn. Representations, pp. 1-26, 2019.
12.
O. Ronneberger, P. Fischer and T. Brox, "U-Net: Convolutional networks for biomedical image segmentation", Proc. Int. Conf. Med. Image Comput. Comput.-Assisted Intervention, pp. 234-241, 2015.
13.
C. Dong, C. C. Loy and X. Tang, "Accelerating the super-resolution convolutional neural network", Proc. Eur. Conf. Comput. Vis., pp. 391-407, 2016.
14.
M. D. Zeiler, D. Krishnan, G. W. Taylor and R. Fergus, "Deconvolutional networks", Proc. IEEE Comput. Soc. Conf. Comput. Vis. Pattern Recognit., pp. 2528-2535, 2010.
15.
W. Shi et al., "Real-time single image and video super-resolution using an efficient sub-pixel convolutional neural network", Proc. IEEE Conf. Comput. Vis. Pattern Recognit., pp. 1874-1883, 2016.
16.
J.-H Kim and J.-S Lee, "Deep residual network with enhanced upscaling module for super-resolution", Proc. IEEE Conf. Comput. Vis. Pattern Recognit. Workshops, pp. 800-808, 2018.
17.
W.-S Lai, J.-B Huang, N. Ahuja and M.-H Yang, "Deep Laplacian pyramid networks for fast and accurate super-resolution", Proc. IEEE Conf. Comput. Vis. Pattern Recognit., pp. 5835-5843, 2017.
18.
M. Haris, G. Shakhnarovich and N. Ukita, "Deep back-projection networks for super-resolution", Proc. Proc. IEEE Conf. Comput. Vis. Pattern Recognit., pp. 1664-1673, 2018.
19.
A. Shocher, N. Cohen and M. Irani, "‘Zero-shot’ super-resolution using deep internal learning", Proc. IEEE Conf. Comput. Vis. Pattern Recognit., pp. 3118-3126, 2018.
20.
D. Ulyanov, A. Vedaldi and V. Lempitsky, "Deep image prior", Proc. IEEE Conf. Comput. Vis. Pattern Recognit., pp. 9446-9454, 2018.
21.
M. Fritsche, S. Gu and R. Timofte, "Frequency separation for real-world super-resolution", Proc. IEEE/CVF Int. Conf. Comput. Vis. Workshop, pp. 3599-3608, 2019.
22.
Y. Zhang, Y. Tian, Y. Kong, B. Zhong and Y. Fu, "Residual dense network for image super-resolution", Proc. IEEE Conf. Comput. Vis. Pattern Recognit., pp. 2472-2481, 2018.
23.
C. Liu et al., "Multi-scale residual hierarchical dense networks for single image super-resolution", IEEE Access, vol. 7, pp. 60572-60583, 2019.
24.
X. Yang et al., "DRFN: Deep recurrent fusion network for single-image super-resolution with large factors", IEEE Trans. Multimedia, vol. 21, no. 2, pp. 328-337, Feb. 2019.
25.
F. Fang, J. Li and T. Zeng, "Soft-edge assisted network for single image super-resolution", IEEE Trans. Image Process., vol. 29, pp. 4656-4668, 2020.
26.
L. Zhang et al., "Adaptive importance learning for improving lightweight image super-resolution network", Int. J. Comput. Vis., vol. 128, no. 2, pp. 479-499, 2020.
27.
T. Dai, J. Cai, Y. Zhang, S.-T Xia and L. Zhang, "Second-order attention network for single image super-resolution", Proc. IEEE Conf. Comput. Vis. Pattern Recognit., pp. 11065-11074, 2019.
28.
Z. Wang et al., "Image quality assessment: From error visibility to structural similarity", IEEE Trans. Image Process., vol. 13, no. 4, pp. 600-612, Apr. 2004.
29.
M. S. Sajjadi, B. Scholkopf and M. Hirsch, "EnhanceNet: Single image super-resolution through automated texture synthesis", Proc. IEEE Int. Conf. Comput. Vis., pp. 4491-4500, 2017.
30.
R. Mechrez, I. Talmi, F. Shama and L. Zelnik-Manor, "Maintaining natural image statistics with the contextual loss", Asian Conf. Comput. Vis., pp. 427-443, 2018.

Contact IEEE to Subscribe

References

References is not available for this document.