TY - JOUR
T1 - Robust Saliency-Driven Quality Adaptation for Mobile 360-Degree Video Streaming
AU - Wang, Shibo
AU - Yang, Shusen
AU - Su, Hairong
AU - Zhao, Cong
AU - Xu, Chenren
AU - Qian, Feng
AU - Wang, Nanbin
AU - Xu, Zongben
N1 - Publisher Copyright:
IEEE
PY - 2024/2/1
Y1 - 2024/2/1
N2 - Mobile 360-degree video streaming has grown significantly in popularity but the quality of experience (QoE) suffers from insufficient and variable wireless network bandwidth. Recently, saliency-driven 360-degree streaming overcomes the buffer size limitation of head movement trajectory (HMT)-driven solutions and thus strikes a better balance between video quality and rebuffering. However, inaccurate network estimations and intrinsic saliency bias still challenge saliency-based streaming approaches, limiting further QoE improvement. To address these challenges, we design a robust saliency-driven quality adaptation algorithm for 360-degree video streaming, RoSal360. Specifically, we present a practical, tile-size-aware deep neural network (DNN) model with a decoupled self-attention architecture to accurately and efficiently predict the transmission time of video tiles. Moreover, we design a reinforcement learning (RL)-driven online correction algorithm to robustly compensate the improper quality allocations due to saliency bias. Through extensive prototype evaluations over real wireless network environments including commodity WiFi, 4 G/LTE, and 5 G links in the wild, RoSal360 significantly enhances the video quality and reduces the rebuffering ratio, thereby improving the viewer QoE, compared to the state-of-the-art algorithms.
AB - Mobile 360-degree video streaming has grown significantly in popularity but the quality of experience (QoE) suffers from insufficient and variable wireless network bandwidth. Recently, saliency-driven 360-degree streaming overcomes the buffer size limitation of head movement trajectory (HMT)-driven solutions and thus strikes a better balance between video quality and rebuffering. However, inaccurate network estimations and intrinsic saliency bias still challenge saliency-based streaming approaches, limiting further QoE improvement. To address these challenges, we design a robust saliency-driven quality adaptation algorithm for 360-degree video streaming, RoSal360. Specifically, we present a practical, tile-size-aware deep neural network (DNN) model with a decoupled self-attention architecture to accurately and efficiently predict the transmission time of video tiles. Moreover, we design a reinforcement learning (RL)-driven online correction algorithm to robustly compensate the improper quality allocations due to saliency bias. Through extensive prototype evaluations over real wireless network environments including commodity WiFi, 4 G/LTE, and 5 G links in the wild, RoSal360 significantly enhances the video quality and reduces the rebuffering ratio, thereby improving the viewer QoE, compared to the state-of-the-art algorithms.
KW - 360-degree video streaming
KW - Quality adaptation
KW - network estimation
KW - saliency
UR - http://www.scopus.com/inward/record.url?scp=85147261932&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85147261932&partnerID=8YFLogxK
U2 - 10.1109/tmc.2023.3235103
DO - 10.1109/tmc.2023.3235103
M3 - Article
AN - SCOPUS:85147261932
SN - 1536-1233
VL - 23
SP - 1312
EP - 1329
JO - IEEE Transactions on Mobile Computing
JF - IEEE Transactions on Mobile Computing
IS - 2
ER -