Attention Down-Sampling Transformer, Relative Ranking and Self-Consistency For Blind Image Quality Assessment | IEEE Conference Publication | IEEE Xplore

Attention Down-Sampling Transformer, Relative Ranking and Self-Consistency For Blind Image Quality Assessment


Abstract:

The no-reference image quality assessment is a challenging domain that addresses estimating image quality without the original reference. We introduce an improved mechani...Show More

Abstract:

The no-reference image quality assessment is a challenging domain that addresses estimating image quality without the original reference. We introduce an improved mechanism to extract local and non-local information from images via different transformer encoders and CNNs. The utilization of Transformer encoders aims to mitigate locality bias and generate a non-local representation by sequentially processing CNN features, which inherently capture local visual structures. Establishing a stronger connection between subjective and objective assessments is achieved through sorting within batches of images based on relative distance information. A self-consistency approach to self-supervision is presented, explicitly addressing the degradation of no-reference image quality assessment (NR-IQA) models under equivariant transformations. Our approach ensures model robustness by maintaining consistency between an image and its horizontally flipped equivalent. Through empirical evaluation of five popular image quality assessment datasets, the proposed model outperforms alternative algorithms in the context of no-reference image quality assessment datasets, especially on smaller datasets. Codes are available at https://github.com/mas94/ADTRS
Date of Conference: 27-30 October 2024
Date Added to IEEE Xplore: 27 September 2024
ISBN Information:

ISSN Information:

Conference Location: Abu Dhabi, United Arab Emirates

Funding Agency:

No metrics found for this document.

1. Introduction

Understanding image quality is essential for many applications; however, it may be difficult since we periodically need an ideal reference image. We can address this issue using Non-Reference Image Quality Assessment (NR-IQA). The objective is to develop techniques that can independently assess the image’s quality without needing the original image. The importance of NR-IQA arises from its wide range of applications, including surveillance systems [1], medical imaging [2], content delivery networks [3], image & video compression [4], etc. It is vital in these domains to assess quality without the original reference image. NR-IQA advances imaging technology and improves user experience. Existing NR-IQA methods focus on developing novel algorithms to handle the problem of evaluating image quality. Test Time Adaptation technique for Image Quality Assessment (TTAIQA) [5], Quality-aware Pre-Trained (QPT) [6] models through self-supervised learning, the Language-Image Quality Evaluator (LIQE), the data-efficient image quality transformer (DEIQT) [7] represents strides in this field and many methods that leverage CNNs. However, shortcomings persist, particularly the limitation imposed by the scarcity of labeled data, hindering the effectiveness of deep learning models and capturing only local features via CNNs while disregarding the nonlocal features of the image that transformers can capture. Popular datasets like the largest NR-IQA dataset, FLIVE, fall short compared to those in other domains, impeding the robust training of NR-IQA models.

Usage
Select a Year
2025

View as

Total usage sinceSep 2024:195
05101520JanFebMarAprMayJunJulAugSepOctNovDec121218000000000
Year Total:42
Data is updated monthly. Usage includes PDF downloads and HTML views.
Contact IEEE to Subscribe

References

References is not available for this document.