Loading web-font TeX/Main/Regular
Towards Neural Codec-Empowered 360- Video Streaming: A Saliency-Aided Synergistic Approach | IEEE Journals & Magazine | IEEE Xplore

Towards Neural Codec-Empowered 360^\circ Video Streaming: A Saliency-Aided Synergistic Approach


Abstract:

Networked 360^\circ video has become increasingly popular. Despite the immersive experience for users, its sheer data volume, even with the latest H.266 coding and view...Show More

Abstract:

Networked 360^\circ video has become increasingly popular. Despite the immersive experience for users, its sheer data volume, even with the latest H.266 coding and viewport adaptation, remains a significant challenge to today's networks. Recent studies have shown that integrating deep learning into video coding can significantly enhance compression efficiency, providing new opportunities for high-quality video streaming. In this work, we conduct a comprehensive analysis of the potential and issues in applying neural codecs to 360^\circ video streaming. We accordingly present \mathsf {NETA}, a synergistic streaming scheme that merges neural compression with traditional coding techniques, seamlessly implemented within an edge intelligence framework. To address the non-trivial challenges in the short viewport prediction window and time-varying viewing directions, we propose implicit-explicit buffer-based prefetching grounded in content visual saliency and bitrate adaptation with smart model switching around viewports. A novel Lyapunov-guided deep reinforcement learning algorithm is developed to maximize user experience and ensure long-term system stability. We further discuss the concerns towards practical development and deployment and have built a working prototype that verifies \mathsf {NETA}’s excellent performance. For instance, it achieves a 27% increment in viewing quality, a 90% reduction in rebuffering time, and a 64% decrease in quality variation on average, compared to state-of-the-art approaches.
Published in: IEEE Transactions on Multimedia ( Volume: 27)
Page(s): 1588 - 1600
Date of Publication: 30 December 2024

ISSN Information:

Funding Agency:


I. Introduction

The 360 videos with Ultra-High-Definition (UHD) have garnered great attention across various applications, including virtual reality, remote education, and entertainment [1], [2], [3], [4], [5]. They offer users immersive visual and acoustic experiences through Head-Mounted Displays (HMDs) such as VIVE Cosmos [6] or Google Cardboard paired with a smartphone [7] However, streaming 360 video poses a significant challenge due to large data volume, requiring stably high network bandwidth, for instance, 200-300Mbps or even higher [1].

Contact IEEE to Subscribe

References

References is not available for this document.