Loading [MathJax]/extensions/MathMenu.js
Learning Spatial Attention for Face Super-Resolution | IEEE Journals & Magazine | IEEE Xplore

Learning Spatial Attention for Face Super-Resolution


Abstract:

General image super-resolution techniques have difficulties in recovering detailed face structures when applying to low resolution face images. Recent deep learning based...Show More

Abstract:

General image super-resolution techniques have difficulties in recovering detailed face structures when applying to low resolution face images. Recent deep learning based methods tailored for face images have achieved improved performance by jointly trained with additional task such as face parsing and landmark prediction. However, multi-task learning requires extra manually labeled data. Besides, most of the existing works can only generate relatively low resolution face images (e.g., 128 × 128), and their applications are therefore limited. In this paper, we introduce a novel SPatial Attention Residual Network (SPARNet) built on our newly proposed Face Attention Units (FAUs) for face super-resolution. Specifically, we introduce a spatial attention mechanism to the vanilla residual blocks. This enables the convolutional layers to adaptively bootstrap features related to the key face structures and pay less attention to those less feature-rich regions. This makes the training more effective and efficient as the key face structures only account for a very small portion of the face image. Visualization of the attention maps shows that our spatial attention network can capture the key face structures well even for very low resolution faces (e.g., 16×16). Quantitative comparisons on various kinds of metrics (including PSNR, SSIM, identity similarity, and landmark detection) demonstrate the superiority of our method over current state-of-the-arts. We further extend SPARNet with multi-scale discriminators, named as SPARNetHD, to produce high resolution results (i.e., 512×512). We show that SPARNetHD trained with synthetic data can not only produce high quality and high resolution outputs for synthetically degraded face images, but also show good generalization ability to real world low quality face images. Codes are available at https://github.com/chaofengc/Face-SPARNet.
Published in: IEEE Transactions on Image Processing ( Volume: 30)
Page(s): 1219 - 1231
Date of Publication: 14 December 2020

ISSN Information:

PubMed ID: 33315560

Funding Agency:

References is not available for this document.

I. Introduction

Face super-resolution (SR), also known as face hallucination, refers to generating high resolution (HR) face images from the corresponding low resolution (LR) inputs. Since there exist many low resolution face images (e.g., faces in surveillance videos) and face analysis algorithms (e.g., face recognition) often perform poorly on such images, there is a growing interest in face SR.

Select All
1.
X. Tang and Z. Li, "Video based face recognition using multiple classifiers", Proc. IEEE Int. Conf. Autom. Face Gesture Recognit., pp. 345-349, May 2004.
2.
X. Tang and Z. Li, "Frame synchronization and multi-level subspace analysis for video based face recognition", Proc. IEEE Conf. Comput. Vis. Pattern Recognit. (CVPR), vol. 2, pp. 2, Jun. 2004.
3.
X. Tang and Z. Li, "Audio-guided video-based face recognition", IEEE Trans. Circuits Syst. Video Technol., vol. 19, no. 7, pp. 955-964, Jul. 2009.
4.
Z. Li, D. Gong, Y. Qiao and D. Tao, "Common feature discriminant analysis for matching infrared face images to optical face images", IEEE Trans. Image Process., vol. 23, no. 6, pp. 2436-2445, Jun. 2014.
5.
D. Gong, Z. Li, J. Liu and Y. Qiao, "Multi-feature canonical correlation analysis for face photo-sketch image retrieval", Proc. ACM Int. Conf. Multimedia, pp. 617-620, 2013.
6.
Z. Li, D. Gong, Q. Li, D. Tao and X. Li, "Mutual component analysis for heterogeneous face recognition", ACM Trans. Intell. Syst. Technol., vol. 7, no. 3, pp. 1-23, Apr. 2016.
7.
Y. Chen, Y. Tai, X. Liu, C. Shen and J. Yang, "FSRNet: End-to-end learning face super-resolution with facial priors", Proc. IEEE Conf. Comput. Vis. Pattern Recognit. (CVPR), pp. 2492-2501, Jun. 2018.
8.
A. Bulat and G. Tzimiropoulos, "Super-FAN: Integrated facial landmark localization and super-resolution of real-world low resolution faces in arbitrary poses with GANs", Proc. IEEE Conf. Comput. Vis. Pattern Recognit. (CVPR), pp. 109-117, Jun. 2018.
9.
X. Yu, B. Fernando, B. Ghanem, F. Porikli and R. Hartley, "Face super-resolution guided by facial component heatmaps", Proc. Eur. Conf. Comput. Vis. (ECCV), pp. 217-233, 2018.
10.
A. Shrivastava, A. Gupta and R. Girshick, "Training region-based object detectors with online hard example mining", Proc. Conf. Comput. Vis. Pattern Recognit. (CVPR), pp. 761-769, 2016.
11.
T.-C. Wang, M.-Y. Liu, J.-Y. Zhu, A. Tao, J. Kautz and B. Catanzaro, "High-resolution image synthesis and semantic manipulation with conditional GANs", Proc. IEEE Conf. Comput. Vis. Pattern Recognit. (CVPR), pp. 8798-8807, Jun. 2018.
12.
S. Baker and T. Kanade, "Limits on super-resolution and how to break them", IEEE Trans. Pattern Anal. Mach. Intell., vol. 24, no. 9, pp. 1167-1183, Sep. 2002.
13.
X. Wang and X. Tang, "Hallucinating face by eigentransformation", IEEE Trans. Syst. Man Cybern. C Appl. Rev., vol. 35, no. 3, pp. 425-434, Aug. 2005.
14.
C. Liu, H.-Y. Shum and W. T. Freeman, "Face hallucination: Theory and practice", Int. J. Comput. Vis., vol. 75, no. 1, pp. 115-134, Jul. 2007.
15.
X. Ma, J. Zhang and C. Qi, "Hallucinating face by position-patch", Pattern Recognit., vol. 43, no. 6, pp. 2224-2236, Jun. 2010.
16.
C. Liu, H.-Y. Shum and C.-S. Zhang, "A two-step approach to hallucinating faces: Global parametric model and local nonparametric model", Proc. IEEE Conf. Comput. Vis. Pattern Recognit. (CVPR), pp. 1, Dec. 2001.
17.
W. Liu, D. Lin and X. Tang, "Hallucinating faces: Tensorpatch super-resolution and coupled residue compensation", Proc. IEEE Conf. Comput. Vis. Pattern Recognit. (CVPR), vol. 2, pp. 478-484, Jun. 2005.
18.
M. F. Tappen and C. Liu, "A Bayesian approach to alignment-based image hallucination", Proc. Eur. Conf. Comput. Vis. (ECCV), pp. 236-249, 2012.
19.
C.-Y. Yang, S. Liu and M.-H. Yang, "Structured face hallucination", Proc. IEEE Conf. Comput. Vis. Pattern Recognit. (CVPR), pp. 1099-1106, Jun. 2013.
20.
Y. Song, J. Zhang, S. He, L. Bao and Q. Yang, "Learning to hallucinate face images via component generation and enhancement", Proc. Int. Joint Conf. Artif. Intell. (IJCAI), pp. 4537-4543, 2017.
21.
S. Zhu, S. Liu, C. C. Loy and X. Tang, "Deep cascaded bi-network for face hallucination", Proc. Eur. Conf. Comput. Vis. (ECCV), pp. 614-630, 2016.
22.
I. Goodfellow et al., "Generative adversarial nets", Proc. Adv. Neural Inf. Process. Syst., pp. 2672-2680, 2014.
23.
X. Yu and F. Porikli, "Face hallucination with tiny unaligned images by transformative discriminative neural networks", Proc. AAAI, vol. 2, pp. 3, 2017.
24.
X. Yu and F. Porikli, "Hallucinating very low-resolution unaligned and noisy face images by transformative discriminative autoencoders", Proc. IEEE Conf. Comput. Vis. Pattern Recognit. (CVPR), pp. 3760-3768, Jul. 2017.
25.
X. Yu, B. Fernando, R. Hartley and F. Porikli, "Super-resolving very low-resolution face images with supplementary attributes", Proc. IEEE/CVF Conf. Comput. Vis. Pattern Recognit. (CVPR), pp. 908-917, Jun. 2018.
26.
K. Zhang et al., "Super-identity convolutional neural network for face hallucination", Proc. Eur. Conf. Comput. Vis. (ECCV), pp. 183-198, 2018.
27.
K. Deokyun, K. Minseon, K. Gihyun and K. Dae-Shik, "Progressive face super-resolution via attention to facial landmark", Proc. 30th Brit. Mach. Vis. Conf. (BMVC), pp. 1-12, 2019.
28.
C. Ma, Z. Jiang, Y. Rao, J. Lu and J. Zhou, "Deep face super-resolution with iterative collaboration between attentive recovery and landmark estimation", Proc. IEEE Conf. Comput. Vis. Pattern Recognit. (CVPR), pp. 5569-5578, Jun. 2020.
29.
V. Mnih et al., "Recurrent models of visual attention", Proc. Adv. Neural Inf. Process. Syst., pp. 2204-2212, 2014.
30.
J. Hu, L. Shen and G. Sun, "Squeeze-and-excitation networks", Proc. IEEE Conf. Comput. Vis. Pattern Recognit. (CVPR), pp. 7132-7141, Jun. 2018.
Contact IEEE to Subscribe

References

References is not available for this document.