I. Introduction
Objective image and video quality assessment (I/VQA) models aim to predict visual quality without the need to collect human subjective scores. These models often rely on statistical regularities (viz., natural scene statistics—NSS) that govern natural images and videos. NSS-derived features may be used to quantify deviations from these statistical properties that are predictive of visual impairments. There are three categories of objective I/VQA models: full-reference (FR), reduced-reference (RR) and no-reference (NR). FR models [1], [2] compare possibly distorted signals against entire reference versions of them. RR models [3] –[7] use only a subset of the reference data to predict quality, whereas NR models use only the distorted image/video to measure quality [8] –[10].