I. Introduction
The latest development of 5G networks and specialized hardware codecs enable 360 video applications to become popular at a staggering speed. Compared with conventional videos, 360 videos are much larger (4× to 6×) due to their panoramic nature. Nevertheless, users can only see the viewport of the 360 videos at every moment. Therefore, delivering the entire 360 video panorama with the same quality brings an inevitable waste of bandwidth and computational resources. Besides, since the motion-to-photon latency requirement of 360 videos (typically 5 41 ms [1]) is much smaller than the Internet request-reply delay (typically 50 400 ms [2]), the client must prefetch the video content users will see in the future to avoid rendering a blank screen. Hence, tile-based viewport-adaptive streaming approaches are proposed [3]. They predict the user’s future viewport and prefetch high-bitrate tiles to this area and low-bitrate tiles to other regions to guarantee the viewport quality while reducing bandwidth waste.