强化学习
计算机科学
吞吐量
架空(工程)
服务质量
功率控制
蜂窝网络
计算机网络
频道(广播)
无线网络
光谱效率
无线
分布式计算
趋同(经济学)
功率(物理)
人工智能
电信
操作系统
物理
量子力学
经济
经济增长
作者
Zhenfeng Sun,Mohammad Reza Nakhai
标识
DOI:10.1109/icc42927.2021.9501055
摘要
In this paper, we address the problem of Device-to-Device communication (D2D) in the next generations of cellular networks, where the number of D2D pairs can grow large and, hence, improving spectral efficiency becomes a crucial design factor. More specifically, decentralized channel selection and power control by D2D pairs for interference mitigation without inflicting a heavy controlling overhead on the network become significant challenges in allocating resources. To this end, we introduce an online distributed reinforcement learning algorithm at D2D pairs to maximize network throughput, while guaranteeing both D2D users’ and cellular users’ (CUs) Quality of Service (QoS) under the dynamic wireless channel environment. To track and evaluate the performance of the proposed online algorithm, we define three metrics, i.e., D2D collision probability, D2D access rate and time-average network throughput. The simulation results confirm the convergence property of the proposed algorithm and shows improved performance in terms of three defined metrics as compared to the celebrated Q-learning-based method.
科研通智能强力驱动
Strongly Powered by AbleSci AI