Loading [MathJax]/extensions/MathMenu.js
PVBLiF: A Pseudo Video-Based Blind Quality Assessment Metric for Light Field Image | IEEE Journals & Magazine | IEEE Xplore

PVBLiF: A Pseudo Video-Based Blind Quality Assessment Metric for Light Field Image


Abstract:

Going beyond traditional 2D imaging is not only an emerging trend of imaging technology, but also the key to a more immersive user experience. Light Field Image (LFI) is ...Show More

Abstract:

Going beyond traditional 2D imaging is not only an emerging trend of imaging technology, but also the key to a more immersive user experience. Light Field Image (LFI) is a typical high-dimensional imaging format, and the quality evaluation of which is very challenging but necessary. In this article, we propose a novel Pseudo Video-based Blind quality assessment metric for Light Field image (PVBLiF). In contrast to most previous Light Field Image Quality Assessment (LF-IQA) metrics, in which different types of 2D representations derived from LFI are used for quality assessment indirectly, our metric exploits a more intuitive 3D representation, named Pseudo Video Block Sequence (PVBS), to evaluate the perceptual quality of LFI. For this purpose, we first divide the LFI into a massive number of non-overlapping PVBSs, which simultaneously contain spatial and angular information of LFI. Then, we propose a novel network (named PVBSNet) based on Convolutional Neural Networks (CNNs) to extract the spatio-angular features of PVBS and further evaluate the PVBS quality. The proposed PVBSNet consists of four stages: multi-information division, intra-feature extraction, cross-feature fusion, and quality regression. Finally, a Saliency- and Variance-guided Pooling (SVPooling) method is presented to integrate all the PVBS quality into the overall quality of LFI. The proposed PVBLiF metric has been extensively evaluated on three widely-used LFI datasets: Win5-LID, NBU-LF1.0, and SHU. Experimental results demonstrate that our proposed PVBLiF metric outperforms state-of-the-art metrics and is capable of highly approximating the performance of human observers.
Published in: IEEE Journal of Selected Topics in Signal Processing ( Volume: 17, Issue: 6, November 2023)
Page(s): 1193 - 1207
Date of Publication: 22 May 2023

ISSN Information:

Funding Agency:

References is not available for this document.

I. Introduction

Immersive imaging technologies, including light field, 360-degree panoramic, and volumetric images/videos, aim to increase the audience presence and improve the immersive visualization, which is difficult to be achieved with traditional 2D imaging. With the recent availability of hand-held light field cameras, Light Field Image (LFI) has received extensive attention from both academia and industry, further offering the possibility for a wide range of applications. Theoretically, LFI records all the information of light rays as they travel in free space, which was first defined as a 7D plenoptic function [1], [2] and further predigested to a 4D model [3] by assuming that light is wavelength- and time-invariant and unobstructed. As a result, LFI is described via a biplane parameterization , where denote the angular coordinates and denote the spatial coordinates. However, despite a series of simplifications, LFI is still very complicated, with inherently high-dimensional characteristics different from traditional 2D images.

Select All
1.
G. Arun, "The light field", J. Math. Phys., vol. 18, pp. 51-151, 1936.
2.
E. H. Adelson and J. R. Bergen, "The plenoptic function and the elements of early vision", Proc. Comput. Models Vis. Process, pp. 3-20, 1991.
3.
M. Levoy and P. Hanrahan, "Light field rendering", Proc. Annu. Conf. Comput. Graph. Interact. Techn., pp. 31-42, 1996.
4.
Y. Liu, L. Fang, D. Gutierrez, Q. Wang, J. Yu and F. Wu, "Introduction to the issue on light field image processing", IEEE J. Sel. Topics Signal Process., vol. 11, no. 7, pp. 923-925, Oct. 2017.
5.
G. Wu et al., "Light field image processing: An overview", IEEE J. Sel. Topics Signal Process., vol. 11, no. 7, pp. 926-954, Oct. 2017.
6.
X. Min, J. Zhou, G. Zhai, P. L. Callet, X. Yang and X. Guan, "A metric for light field reconstruction compression and display quality evaluation", IEEE Trans. Image Process., vol. 29, pp. 3790-3804, 2020.
7.
C. Meng, P. An, X. Huang, C. Yang, L. Shen and B. Wang, "Objective quality assessment of lenslet light field image based on focus stack", IEEE Trans. Multimedia, vol. 24, pp. 3193-3207, 2022.
8.
W. Zhou, L. Shi, Z. Chen and J. Zhang, "Tensor oriented no-reference light field image quality assessment", IEEE Trans. Image Process., vol. 29, pp. 4070-4084, 2020.
9.
Q. Li and Z. Wang, "Reduced-reference image quality assessment using divisive normalization-based image representation", IEEE J. Sel. Topics Signal Process., vol. 3, no. 2, pp. 202-211, Apr. 2009.
10.
V. Laparra, J. Ballé, A. Berardino and E. P. Simoncelli, "Perceptual image quality assessment using a normalized Laplacian pyramid", Electron. Imag., no. 16, pp. 1-6, 2016.
11.
P. G. Freitas, W. Y. L. Akamine and M. C. Q. Farias, "No-reference image quality assessment based on statistics of local ternary pattern", Proc. IEEE Int. Conf. Qual. Multimedia Experience, pp. 1-6, 2016.
12.
P. G. Freitas, W. Y. L. Akamine and M. C. Q. Farias, "No-reference image quality assessment using orthogonal color planes patterns", IEEE Trans. Multimedia, vol. 20, pp. 3353-3360, 2018.
13.
T. Chinen et al., "Towards a semantic perceptual image metric", Proc. IEEE Int. Conf. Image Process., pp. 624-628, 2018.
14.
Z. Zhou, J. Li, Y. Quan and R. Xu, "Image quality assessment using kernel sparse coding", IEEE Trans. Multimedia, vol. 23, pp. 1592-1604, 2021.
15.
X. Xu, Y. Zhao and Y. Dong, "No-reference stereoscopic image quality assessment based on saliency-guided binocular feature consolidation", Electron. Lett., vol. 53, no. 22, pp. 1468-1470, 2017.
16.
L. Liu, B. Liu, C.-C. Su, H. Huang and A. C. Bovik, "Binocular spatial activity and reverse saliency driven no-reference stereopair quality assessment", Signal Process. Image Commun., vol. 58, pp. 287-299, 2017.
17.
Z. Chen, W. Zhou and W. Li, "Blind stereoscopic video quality assessment: From depth perception to overall experience", IEEE Trans. Image Process., vol. 27, no. 2, pp. 721-734, Feb. 2018.
18.
S. Ling, J. Li, Z. Che, X. Min, G. Zhai and P. Le Callet, "Quality assessment of free-viewpoint videos by quantifying the elastic changes of multi-scale motion trajectories", IEEE Trans. Image Process., vol. 30, pp. 517-531, 2021.
19.
L. Shi, S. Zhao, W. Zhou and Z. Chen, "Perceptual evaluation of light field image", Proc. IEEE 25th Int. Conf. Image Process., pp. 41-45, 2018.
20.
F. Dai, J. Zhang, Y. Ma and Y. Zhang, "Lenselet image compression scheme based on subaperture images streaming", Proc. IEEE Int. Conf. Image Process., pp. 4733-4737, 2015.
21.
D. Liu, L. Wang, L. Li, Z. Xiong, F. Wu and W. Zeng, "Pseudo-sequence-based light field image compression", Proc. IEEE Int. Conf. Multimedia Expo Workshops, pp. 1-4, 2016.
22.
G. Wang, W. Xiang, M. Pickering and C. W. Chen, "Light field multi-view video coding with two-directional parallel inter-view prediction", IEEE Trans. Image Process., vol. 25, no. 11, pp. 5104-5117, Nov. 2016.
23.
L. Li, Z. Li, B. Li, D. Liu and H. Li, "Pseudo-sequence-based 2-D hierarchical coding structure for light-field image compression", IEEE J. Sel. Topics Signal Process., vol. 11, no. 7, pp. 1107-1119, Oct. 2017.
24.
H. Amirpour, M. Pereira and A. Pinheiro, "High efficient snake order pseudo-sequence based light field image compression", Proc. IEEE Data Compress. Conf., pp. 397, 2018.
25.
Z. Huang, M. Yu, G. Jiang, K. Chen, Z. Peng and F. Chen, "Reconstruction distortion oriented light field image dataset for visual communication", Proc. IEEE Int. Symp. Netw. Comput. Commun., pp. 1-5, 2019.
26.
L. Shan, P. An, C. Meng, X. Huang, C. Yang and L. Shen, "A no-reference image quality assessment metric by multiple characteristics of light field images", IEEE Access, vol. 7, pp. 127217-127229, 2019.
27.
A. Zizien and K. Fliegel, "LFDD: Light field image dataset for performance evaluation of objective quality metrics", Proc. SPIE, vol. 11510, 2020.
28.
J. Wu, Y. Liu, W. Dong, G. Shi and W. Lin, "Quality assessment for video with degradation along salient trajectories", IEEE Trans. Multimedia, vol. 21, pp. 2738-2749, 2019.
29.
P. Chen, L. Li, L. Ma, J. Wu and G. Shi, "RirNet: Recurrent-in-recurrent network for video quality assessment", Proc. ACM Int. Conf. Multimedia, pp. 834-842, 2020.
30.
M. Xu, J. Chen, H. Wang, S. Liu, G. Li and Z. Bai, "C3DVQA: Full-reference video quality assessment with 3D convolutional neural network", Proc. IEEE Int. Conf. Acoust. Speech Signal Process., pp. 4447-4451, 2020.
Contact IEEE to Subscribe

References

References is not available for this document.