IEICE Transactions on Communications
Online ISSN : 1745-1345
Print ISSN : 0916-8516

この記事には本公開記事があります。本公開記事を参照してください。
引用する場合も本公開記事を引用してください。

Reinforcement Learning for QoS-Constrained Autonomous Resource Allocation with H2H/M2M Co-Existence in Cellular Networks
Xing WEIXuehua LIShuo CHENNa LI
著者情報
ジャーナル 認証あり 早期公開

論文ID: 2021TMP0011

この記事には本公開記事があります。
詳細
抄録

Machine-to-Machine (M2M) communication plays a pivotal role in the evolution of Internet of Things (IoT). Cellular networks are considered to be a key enabler for M2M communications, which are originally designed mainly for Human-to-Human (H2H) communications. The introduction of M2M users will cause a series of problems to traditional H2H users, i.e., interference between various traffic. Resource allocation is an effective solution to these problems. In this paper, we consider a shared resource block (RB) and power allocation in an H2H/M2M coexistence scenario, where M2M users are subdivided into delay-tolerant and delay-sensitive types. We first model the RB-power allocation problem as maximization of capacity under Quality-of-Service (QoS) constraints of different types of traffic. Then, a learning framework is introduced, wherein a complex agent is built from simpler subagents, which provides the basis for distributed deployment scheme. Further, we proposed distributed Q-learning based autonomous RB-power allocation algorithm (DQ-ARPA), which enables the machine type network gateways (MTCG) as agents to learn the wireless environment and choose the RB-power autonomously to maximize M2M pairs' capacity while ensuring the QoS requirements of critical services. Simulation results indicates that with an appropriate reward design, our proposed scheme succeeds in reducing the impact of delay-tolerant machine type users on critical services in terms of SINR thresholds and outage ratios.

著者関連情報
© 2022 The Institute of Electronics, Information and Communication Engineers
feedback
Top