Spatial and Temporal Consistency-Aware Dynamic Adaptive Streaming for 360-Degree Videos

Research output: Journal Publications and ReviewsRGC 21 - Publication in refereed journalpeer-review

26 Scopus Citations
View graph of relations


Related Research Unit(s)


Original languageEnglish
Pages (from-to)177-193
Journal / PublicationIEEE Journal on Selected Topics in Signal Processing
Issue number1
Online published6 Dec 2019
Publication statusPublished - Jan 2020


The 360-degree video allows users to enjoy the whole scene by interactively switching viewports. However, the huge data volume of the 360-degree video limits its remote applications via network. To provide high quality of experience (QoE) for remote web users, this paper presents a tile-based adaptive streaming method for 360-degree videos. First, we propose a simple yet effective rate adaptation algorithm to determine the requested bitrate for downloading the current video segment by considering the balance between the buffer length and video quality. Then, we propose to use a Gaussian model to predict the field of view at the beginning of each requested video segment. To deal with the circumstance that the view angle is switched during the display of a video segment, we propose to download all the tiles in the 360-degree video with different priorities based on a Zipf model. Finally, in order to allocate bitrates for all the tiles, a two-stage optimization algorithm is proposed to preserve the quality of tiles in FoV and guarantee the spatial and temporal smoothness. Experimental results demonstrate the effectiveness and advantage of the proposed method compared with the state-of-the-art methods. That is, our method preserves both the quality and the smoothness of tiles in FoV, thus providing the best QoE for users.

Research Area(s)

  • 360-degree video, field of view, rate adaptation, DASH, video compression, quality of experience