Loading [MathJax]/extensions/MathMenu.js
Swift Parameter-free Attention Network for Efficient Super-Resolution | IEEE Conference Publication | IEEE Xplore

Swift Parameter-free Attention Network for Efficient Super-Resolution


Abstract:

Single Image Super-Resolution (SISR) is a crucial task in low-level computer vision, aiming to reconstruct high-resolution images from low-resolution counterparts. Conven...Show More

Abstract:

Single Image Super-Resolution (SISR) is a crucial task in low-level computer vision, aiming to reconstruct high-resolution images from low-resolution counterparts. Conventional attention mechanisms have significantly improved SISR performance but often result in complex network structures and large number of parameters, leading to slow inference speed and large model size. To address this issue, we propose the Swift Parameter-free Attention Network (SPAN), a highly efficient SISR model that balances parameter count, inference speed, and image quality. SPAN employs a novel parameter-free attention mechanism, which leverages symmetric activation functions and residual connections to enhance high-contribution information and suppress redundant information. Our theoretical analysis demonstrates the effectiveness of this design in achieving the attention mechanism’s purpose. We evaluate SPAN on multiple benchmarks, showing that it outperforms existing efficient super-resolution models in terms of both image quality and inference speed, achieving a significant quality-speed trade-off. This makes SPAN highly suitable for real-world applications, particularly in resource-constrained scenarios. Notably, we won the first place both in the overall performance track and runtime track of the NTIRE 2024 efficient super-resolution challenge. Our code and models are made publicly available at https://github.com/hongyuanyu/span.
Date of Conference: 17-18 June 2024
Date Added to IEEE Xplore: 27 September 2024
ISBN Information:

ISSN Information:

Conference Location: Seattle, WA, USA
References is not available for this document.

1. Introduction

Single Image Super-Resolution (SISR) is a well-established task in low-level computer vision, which aims to reconstruct a high-resolution image from a single low-resolution image. This task has broad applicability in enhancing image quality across various domains [16], [37], [43], [44], [48], [49], [57]. The advent of deep learning has led to significant advancements in this field [2], [10], [12], [19], [24], [32], [34], [36], [50], [59]. Recent progress in super-resolution tasks has been largely driven by the attention mechanism. Numerous state-of-the-art super-resolution networks incorporate attention mechanisms or even employ larger vision transformers (ViTs) as the model architecture [6], [8], [20], [27], [32], [35], [42], [53], [60]. These networks emphasize key features and long-distance dependencies between patches through attention maps, capturing a wider range of contextual information to ensure continuity of details and accuracy of edge textures. However, the computational requirements of the attention mechanism, which involve complex network structures and a substantial number of additional parameters, lead to challenges such as large model size and slow inference speed. These challenges limit the applicability of these models, hindering their use in efficient, high-speed computing scenarios, such as SISR tasks on resource-constrained mobile devices.

Select All
1.
Eirikur Agustsson and Radu Timofte, "Ntire 2017 challenge on single image super-resolution: Dataset and study", The IEEE Conference on Computer Vision and Pattern Recognition (CVPR) Workshops, 2017.
2.
Namhyuk Ahn, Byungkon Kang and Kyung-Ah Sohn, "Fast accurate and lightweight super-resolution with cascading residual network", Computer Vision - ECCV 2018 - 15th European Conference Munich Germany September 8-14 2018 Proceedings Part X, pp. 256-272, 2018.
3.
Namhyuk Ahn, Byungkon Kang and Kyung-Ah Sohn, "Fast accurate and lightweight super-resolution with cascading residual network", Proceedings of the European conference on computer vision (ECCV), pp. 252-268, 2018.
4.
Marco Bevilacqua, Aline Roumy, Christine Guillemot and Marie Line Alberi-Morel, "Low-complexity single-image super-resolution based on nonnegative neighbor embedding", 2012.
5.
Chenjie Cao, Qiaole Dong and Yanwei Fu, "Learning prior feature and attention enhanced image inpainting", European Conference on Computer Vision, pp. 306-322, 2022.
6.
Jiezhang Cao, Qin Wang, Yongqin Xian, Yawei Li, Bingbing Ni, Zhiming Pi, et al., "Ciaosr: Continuous implicit attention-in-attention network for arbitrary-scale image super-resolution", Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 1796-1807, 2023.
7.
Junsuk Choe and Hyunjung Shim, "Attention-based dropout layer for weakly supervised object localization", Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 2219-2228, 2019.
8.
Haram Choi, Jeongmin Lee and Jihoon Yang, "N-gram in swin transformers for efficient lightweight image super-resolution", Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 2071-2081, 2023.
9.
Xiaohan Ding, Xiangyu Zhang, Ningning Ma, Jungong Han, Guiguang Ding and Jian Sun, "Repvgg: Making vgg-style convnets great again", Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 13733-13742, 2021.
10.
Chao Dong, Chen Change Loy, Kaiming He and Xiaoou Tang, "Learning a deep convolutional network for image super-resolution", Computer Vision - ECCV 2014 - 13th European Conference Zurich Switzerland September 6-12 2014 Proceedings Part IV, pp. 184-199, 2014.
11.
Chao Dong, Chen Change Loy, Kaiming He and Xiaoou Tang, "Learning a deep convolutional network for image super-resolution", Computer Vision–ECCV 2014: 13th European Conference Zurich Switzerland September 6-12 2014 Proceedings Part IV 13, pp. 184-199, 2014.
12.
Chao Dong, Chen Change Loy and Xiaoou Tang, "Accelerating the super-resolution convolutional neural network", Computer Vision - ECCV 2016 - 14th European Conference Amsterdam The Netherlands October 11-14 2016 Proceedings Part II, pp. 391-407, 2016.
13.
Jie Du, Kai Guan, Yanhong Zhou, Yuanman Li and Tianfu Wang, "Parameter-free similarity-aware attention module for medical image classification and segmentation", IEEE Transactions on Emerging Topics in Computational Intelligence, 2022.
14.
Guangwei Gao, Wenjie Li, Juncheng Li, Fei Wu, Huimin Lu and Yi Yu, "Feature distillation interaction weighting network for lightweight image super-resolution", CoRR, 2021.
15.
Shenyuan Gao, Chunluan Zhou, Chao Ma, Xinggang Wang and Junsong Yuan, "Aiatrack: Attention in attention for transformer visual tracking", European Conference on Computer Vision, pp. 146-164, 2022.
16.
C Heltin Genitha and K Vani, "Super resolution mapping of satellite images using hopfield neural networks" in Recent Advances in Space Technology Services and Climate Change 2010 (RSTS & CC-2010), IEEE, pp. 114-118, 2010.
17.
Daniel Haase and Manuel Amthor, "Rethinking depthwise separable convolutions: How intra-kernel correlations lead to improved mobilenets", Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 14600-14609, 2020.
18.
Jia-Bin Huang, Abhishek Singh and Narendra Ahuja, "Single image super-resolution from transformed self-exemplars", Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 5197-5206, 2015.
19.
Zheng Hui, Xinbo Gao, Yunchu Yang and Xiumei Wang, "Lightweight image super-resolution with information multi-distillation network", Proceedings of the 27th ACM International Conference on Multimedia MM 2019 Nice France October 21-25 2019, pp. 2024-2032, 2019.
20.
Zheng Hui, Xinbo Gao, Yunchu Yang and Xiumei Wang, "Lightweight image super-resolution with information multi-distillation network", Proceedings of the 27th acm international conference on multimedia, pp. 2024-2032, 2019.
21.
Carlo Innamorati, Tobias Ritschel, Tim Weyrich and Niloy J Mitra, "Learning on the edge: Investigating boundary filters in cnns", International Journal of Computer Vision, vol. 128, pp. 773-782, 2020.
22.
Younho Jang, Wheemyung Shin, Jinbeom Kim, Simon Woo and Sung-Ho Bae, "Glamd: Global and local attention mask distillation for object detectors", European Conference on Computer Vision, pp. 460-476, 2022.
23.
Jiwon Kim, Jung Kwon Lee and Kyoung Mu Lee, "Deeply-recursive convolutional network for image super-resolution", 2016 IEEE Conference on Computer Vision and Pattern Recognition CVPR 2016 Las Vegas NV USA June 27-30 2016, pp. 1637-1645, 2016.
24.
Jiwon Kim, Jung Kwon Lee and Kyoung Mu Lee, "Accurate image super-resolution using very deep convolutional networks", 2016 IEEE Conference on Computer Vision and Pattern Recognition CVPR 2016 Las Vegas NV USA June 27-30 2016, pp. 1646-1654, 2016.
25.
Jiwon Kim, Jung Kwon Lee and Kyoung Mu Lee, "Deeplyrecursive convolutional network for image super-resolution", Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 1637-1645, 2016.
26.
Diederik P Kingma and Jimmy Ba, "Adam: A method for stochastic optimization", 2014.
27.
Fangyuan Kong, Mingxi Li, Songwei Liu, Ding Liu, Jingwen He, Yang Bai, et al., "Residual local feature network for efficient super-resolution", Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 766-776, 2022.
28.
Wei-Sheng Lai, Jia-Bin Huang, Narendra Ahuja and Ming-Hsuan Yang, "Deep laplacian pyramid networks for fast and accurate super-resolution", 2017 IEEE Conference on Computer Vision and Pattern Recognition CVPR 2017 Honolulu HI USA July 21-26 2017, pp. 5835-5843, 2017.
29.
Hunsang Lee, Hyesong Choi, Kwanghoon Sohn and Dongbo Min, "Knn local attention for image restoration", Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 2139-2149, 2022.
30.
Wenbo Li, Kun Zhou, Lu Qi, Nianjuan Jiang, Jiangbo Lu and Jiaya Jia, "Lapar: Linearly-assembled pixel-adaptive regression network for single image super-resolution and beyond", NeurIPS, pp. 20343-20355, 2020.
Contact IEEE to Subscribe

References

References is not available for this document.