EffiHDR: An Efficient Framework for HDRTV Reconstruction and Enhancement in UHD Systems | IEEE Journals & Magazine | IEEE Xplore

EffiHDR: An Efficient Framework for HDRTV Reconstruction and Enhancement in UHD Systems


Abstract:

Recent advancements in SDRTV-to-HDRTV conversion have yielded impressive results in reconstructing high dynamic range television (HDRTV) videos from standard dynamic rang...Show More

Abstract:

Recent advancements in SDRTV-to-HDRTV conversion have yielded impressive results in reconstructing high dynamic range television (HDRTV) videos from standard dynamic range television (SDRTV) videos. However, the practical applications of these techniques are limited for ultra-high definition (UHD) video systems due to their high computational and memory costs. In this paper, we propose EffiHDR, an efficient framework primarily operating in the downsampled space, effectively reducing the computational and memory demands. Our framework comprises a real-time SDRTV-to-HDRTV Reconstruction model and a plug-and-play HDRTV Enhancement model. The SDRTV-to-HDRTV Reconstruction model learns affine transformation coefficients instead of directly predicting output pixels to preserve high-frequency information and mitigate information loss caused by downsampling. It decomposes SDRTV-to-HDR mapping into pixel intensity-dependent and local-dependent affine transformations. The pixel intensity-dependent transformation leverages global contexts and pixel intensity conditions to transform SDRTV pixels to the HDRTV domain. The local-dependent transformation predicts affine coefficients based on local contexts, further enhancing dynamic range, local contrast, and color tone. Additionally, we introduce a plug-and-play HDRTV Enhancement model based on an efficient Transformer-based U-net, which enhances luminance and color details in challenging recovery scenarios. Experimental results demonstrate that our SDRTV-to-HDRTV Reconstruction model achieves real-time 4K conversion with impressive performance. When combined with the HDRTV Enhancement model, our approach outperforms state-of-the-art methods in performance and efficiency.
Published in: IEEE Transactions on Broadcasting ( Volume: 70, Issue: 2, June 2024)
Page(s): 620 - 636
Date of Publication: 10 January 2024

ISSN Information:

Funding Agency:

References is not available for this document.

I. Introduction

High dynamic range television (HDRTV1) videos have significantly improved the visual quality. Compared to standard dynamic range television (SDRTV) videos, HDRTV videos offer a broader range of brightness levels and a wider color gamut, enabling a more realistic representation of the natural world. With the increasing support for HDR displays in devices like televisions, smartphones, and computers, there is a growing demand for converting existing SDRTV videos to HDRTV in order to enhance video broadcasting services. In this paper, we refer to the task of converting SDRTV videos to HDRTV videos as SDRTV-to-HDRTV, adopting the definition provided by HDRTVNet [1]. It is essential to note that the intent of SDRTV-to-HDRTV differs from Inverse Tone Mapping (ITM), as illustrated in Fig. 2. ITM aims to restore the original luminance of the captured scene, focusing on recovering information in saturated areas. In contrast, SDRTV-to-HDRTV aims to improve the visual quality of video on HDR displays, offering comprehensive enhancements in dynamic range, gamut, and local contrast.

We add a suffix TV after HDR to indicate content in HDR-TV format and standard(HDR10 and Dolby Vision, etc.).

Efficiency comparison between our method and other works. EffiHDR-Base is our SDRTV-to-HDRTV Reconstruction model with the downsampling factor 8. EffiHDR-Enhancer is the cascade of our EffiHDR-Base model and HDRTV Enhancement model.

SDRTV-to-HDRTV differs from Inverse Tone Mapping (ITM) functionally. ITM focuses on restoring the original luminance of the captured scene. To display predicted HDR contents, tone mapping may be needed to compress the dynamic range. Because the dynamic range that HDR displays can present is still limited. In contrast, SDRTV-to-HDRTV aims to directly predict the HDRTV videos for HDR display.

Select All
1.
X. Chen, Z. Zhang, J. S. Ren, L. Tian, Y. Qiao and C. Dong, "A new journey from SDRTV to HDRTV", Proc. ICCV, pp. 4480-4489, 2021.
2.
S. Y. Kim, J. Oh and M. Kim, "Deep SR-ITM: Joint learning of super-resolution and inverse tone-mapping for 4K UHD HDR applications", Proc. ICCV, pp. 3116-3125, 2019.
3.
S. Y. Kim, J. Oh and M. Kim, "JSI-GAN: GAN-based joint super-resolution and inverse tone-mapping with pixel-wise task-specific filters for UHD HDR video", Proc. AAAI, pp. 11287-11295, 2020.
4.
G. He, S. Long, L. Xu, C. Wu, W. Yu and J. Zhou, "Global priors guided modulation network for joint super-resolution and SDRTV-to-HDRTV", Neurocomputing, vol. 554, Oct. 2023.
5.
H. Zhang, L. Song, W. Gan and R. Xie, "Multi-scale-based joint super-resolution and inverse tone-mapping with data synthesis for UHD HDR video", Displays, vol. 79, Sep. 2023.
6.
G. He et al., "SDRTV-to-HDRTV via hierarchical dynamic context feature mapping", Proc. ACM Multimedia, pp. 2890-2898, 2022.
7.
G. Xu, Q. Hou, L. Zhang and M. Cheng, "FMNet: Frequency-aware modulation network for SDR-to-HDR translation", Proc. ACM Multimedia, pp. 6425-6435, 2022.
8.
C. Guo, L. Fan, Z. Xue and X. Jiang, "Learning a practical SDR-to-HDRTV up-conversion using new dataset and degradation models", Proc. CVPR, pp. 22231-22241, 2023.
9.
K. Xu et al., "Towards robust SDRTV-to-HDRTV via dual inverse degradation network", arXiv:2307.03394, 2023.
10.
P. J. Burt and E. H. Adelson, "The Laplacian pyramid as a compact image code", IEEE Trans. Commun., vol. 31, no. 4, pp. 532-540, Apr. 1983.
11.
Image parameter values for high dynamic range television for use in production and international programme exchange, Geneva, Switzerland:ITU-R, 2018.
12.
Colour conversion from recommendation ITU-R BT.709 to recommendation ITU-R BT.2020, Geneva, Switzerland:ITU-R, 2015.
13.
H. Wu, S. Zheng, J. Zhang and K. Huang, "Fast end-to-end trainable guided filter", Proc. CVPR, pp. 1838-1847, 2018.
14.
Y. Huo, F. Yang, L. Dong and V. Brost, "Physiological inverse tone mapping based on retina response", Vis. Comput., vol. 30, no. 5, pp. 507-517, 2014.
15.
R. P. Kovaleski and M. M. de Oliveira Neto, "High-quality reverse tone mapping for a wide range of exposures", Proc. SIBGRAPI, pp. 49-56, 2014.
16.
P. Mohammadi, M. T. Pourazad and P. Nasiopoulos, "A perception-based inverse tone mapping operator for high dynamic range video applications", IEEE Trans. Circuits Syst. Video Technol., vol. 31, no. 5, pp. 1711-1723, May 2021.
17.
D. Kim and M. Kim, "Learning-based low-complexity reverse tone mapping with linear mapping", IEEE Trans. Circuits Syst. Video Technol., vol. 30, no. 2, pp. 400-414, Feb. 2020.
18.
K. He, J. Sun and X. Tang, "Guided image filtering", IEEE Trans. Pattern Anal. Mach. Intell., vol. 35, no. 6, pp. 1397-1409, Jun. 2013.
19.
Z. Shi et al., "Learning for unconstrained space-time video super-resolution", IEEE Trans. Broadcast., vol. 68, no. 2, pp. 345-358, Jun. 2022.
20.
A. Esmaeilzehi, M. O. Ahmad and M. N. S. Swamy, "SRNMSM: A deep light-weight image super resolution network using multi-scale spatial and morphological feature generating residual blocks", IEEE Trans. Broadcast., vol. 68, no. 1, pp. 58-68, Mar. 2022.
21.
J. Wan, H. Yin, Z. Liu, A. Chong and Y. Liu, "Lightweight image super-resolution by multi-scale aggregation", IEEE Trans. Broadcast., vol. 67, no. 2, pp. 372-382, Jun. 2021.
22.
J. Wan, H. Yin, Z. Liu, Y. Liu and S. Wang, "Multi-stage edge-guided stereo feature interaction network for stereoscopic image super-resolution", IEEE Trans. Broadcast., vol. 69, no. 2, pp. 357-368, Jun. 2023.
23.
P. Isola, J. Zhu, T. Zhou and A. A. Efros, "Image-to-image translation with conditional adversarial networks", Proc. CVPR, pp. 5967-5976, 2017.
24.
M. Gharbi, J. Chen, J. T. Barron, S. W. Hasinoff and F. Durand, "Deep bilateral learning for real-time image enhancement", ACM Trans. Graph., vol. 36, no. 4, pp. 118, 2017.
25.
W. Song, Y. Wang, D. Huang, A. Liotta and C. Perra, "Enhancement of underwater images with statistical model of background light and optimization of transmission map", IEEE Trans. Broadcast., vol. 66, no. 1, pp. 153-169, Mar. 2020.
26.
N. Qiao, L. Dong and C. Sun, "Adaptive deep learning network with multi-scale and multi-dimensional features for underwater image enhancement", IEEE Trans. Broadcast., vol. 69, no. 2, pp. 482-494, Jun. 2023.
27.
Y. Endo, Y. Kanamori and J. Mitani, "Deep reverse tone mapping", ACM Trans. Graph., vol. 36, no. 6, pp. 177, 2017.
28.
G. Eilertsen, J. Kronander, G. Denes, R. K. Mantiuk and J. Unger, "HDR image reconstruction from a single exposure using deep CNNs", ACM Trans. Graph., vol. 36, no. 6, pp. 178, 2017.
29.
D. Marnerides, T. Bashford-Rogers, J. Hatchett and K. Debattista, "ExpandNet: A deep convolutional neural network for high dynamic range expansion from low dynamic range content", Comput. Graph. Forum, vol. 37, no. 2, pp. 37-49, 2018.
30.
Y. Xu, S. Ning, R. Xie and L. Song, "Gan based multi-exposure inverse tone mapping", Proc. ICIP, pp. 4365-4369, 2019.
Contact IEEE to Subscribe

References

References is not available for this document.