TY - JOUR
T1 - Multi-User Adaptive Video Delivery over Wireless Networks
T2 - A Physical Layer Resource-Aware Deep Reinforcement Learning Approach
AU - Tang, Kexin
AU - Kan, Nuowen
AU - Zou, Junni
AU - Li, Chenglin
AU - Fu, Xiao
AU - Hong, Mingyi
AU - Xiong, Hongkai
N1 - Publisher Copyright:
© 1991-2012 IEEE.
PY - 2021/2/1
Y1 - 2021/2/1
N2 - In this paper, we investigate the adaptive video delivery for multiple users over time-varying and mutually interfering multi-cell wireless networks. The key research challenge is to jointly design the physical-layer resource allocation scheme and application-layer rate adaptation logic, such that the users' long-term fair quality of experience (QoE) can be maximized. Due to the timescale mismatch between these two layers and the asynchrony of user requests, however, it is difficult to directly model the cross-layer stochastic control problem by using a reinforcement learning framework. To address this difficulty, we propose a novel two-level decision framework where an optimization-based beamforming scheme (performed at the base stations) and a deep reinforcement learning (DRL)-based rate adaptation scheme (performed at the user terminals) are, respectively, developed, such that a highly complex long-term multi-user QoE fairness problem is decomposed into some relatively simple problems and solved effectively. Our strategy represents a significant departure from the existing schemes with consideration of either a short-term multi-user QoE maximization or a long-term single-user point-to-point QoE maximization. Extensive simulations demonstrate that the proposed cross-layer design is effective and promising.
AB - In this paper, we investigate the adaptive video delivery for multiple users over time-varying and mutually interfering multi-cell wireless networks. The key research challenge is to jointly design the physical-layer resource allocation scheme and application-layer rate adaptation logic, such that the users' long-term fair quality of experience (QoE) can be maximized. Due to the timescale mismatch between these two layers and the asynchrony of user requests, however, it is difficult to directly model the cross-layer stochastic control problem by using a reinforcement learning framework. To address this difficulty, we propose a novel two-level decision framework where an optimization-based beamforming scheme (performed at the base stations) and a deep reinforcement learning (DRL)-based rate adaptation scheme (performed at the user terminals) are, respectively, developed, such that a highly complex long-term multi-user QoE fairness problem is decomposed into some relatively simple problems and solved effectively. Our strategy represents a significant departure from the existing schemes with consideration of either a short-term multi-user QoE maximization or a long-term single-user point-to-point QoE maximization. Extensive simulations demonstrate that the proposed cross-layer design is effective and promising.
KW - Wireless video streaming
KW - beamforming
KW - cross-layer design
KW - rate adaptation
KW - reinforcement learning
UR - https://www.scopus.com/pages/publications/85100550480
UR - https://www.scopus.com/pages/publications/85100550480#tab=citedBy
U2 - 10.1109/tcsvt.2020.2980587
DO - 10.1109/tcsvt.2020.2980587
M3 - Article
AN - SCOPUS:85100550480
SN - 1051-8215
VL - 31
SP - 798
EP - 815
JO - IEEE Transactions on Circuits and Systems for Video Technology
JF - IEEE Transactions on Circuits and Systems for Video Technology
IS - 2
M1 - 9035396
ER -