Access and Radio Resource Management for IAB Networks Using Deep Reinforcement Learning
Congestion in dense traffic networks is a prominent obstacle towards realizing the performance requirements of 5G new radio. Since traditional adaptive traffic signal control cannot resolve this type of congestion, realizing context in the network and adapting resource allocation based on real-time...
Main Authors: | , , |
---|---|
Format: | Article |
Language: | English |
Published: |
IEEE
2021-01-01
|
Series: | IEEE Access |
Subjects: | |
Online Access: | https://ieeexplore.ieee.org/document/9512096/ |
id |
doaj-c3e07ecd2cad43b098dd2258f5a1fd13 |
---|---|
record_format |
Article |
spelling |
doaj-c3e07ecd2cad43b098dd2258f5a1fd132021-08-20T23:00:23ZengIEEEIEEE Access2169-35362021-01-01911421811423410.1109/ACCESS.2021.31043229512096Access and Radio Resource Management for IAB Networks Using Deep Reinforcement LearningMalcolm M. Sande0https://orcid.org/0000-0002-6636-2050Mduduzi C. Hlophe1https://orcid.org/0000-0001-6111-5619Bodhaswar T. Maharaj2https://orcid.org/0000-0002-3703-3637Department of Electrical, Electronic and Computer Engineering, University of Pretoria, Pretoria, South AfricaDepartment of Electrical, Electronic and Computer Engineering, University of Pretoria, Pretoria, South AfricaDepartment of Electrical, Electronic and Computer Engineering, University of Pretoria, Pretoria, South AfricaCongestion in dense traffic networks is a prominent obstacle towards realizing the performance requirements of 5G new radio. Since traditional adaptive traffic signal control cannot resolve this type of congestion, realizing context in the network and adapting resource allocation based on real-time parameters is an attractive approach. This article proposes a radio resource management solution for congestion avoidance on the access side of an integrated access and backhaul (IAB) network using deep reinforcement learning (DRL). The objective of this article is to obtain an optimal policy under which the transmission throughput of all UEs is maximized under the dictates of environmental pressures such as traffic load and transmission power. Here, the resource management problem was converted into a constrained problem using Markov decision processes and dynamic power management, where a deep neural network was trained for optimal power allocation. By initializing a power control parameter, <inline-formula> <tex-math notation="LaTeX">$\theta _{t}$ </tex-math></inline-formula>, with zero-mean normal distribution, the DRL algorithm adopts a learning policy that aims to achieve logical allocation of resources by placing more emphasis on congestion control and user satisfaction. The performance of the proposed DRL algorithm was evaluated using two learning schemes, i.e., individual learning and nearest neighbor cooperative learning, and this was compared with the performance of a baseline algorithm. The simulation results indicate that the proposed algorithms give better overall performance when compared to the baseline algorithm. From the simulation results, there is a subtle, but critically important insight that brings into focus the fundamental connection between learning rate and the two proposed algorithms. The nearest neighbor cooperative learning algorithm is suitable for IAB networks because its throughput has a good correlation with the congestion rate.https://ieeexplore.ieee.org/document/9512096/Congestion controldeep reinforcement learningintegrated access and backhaulmillimeter wavenearest neighborresource allocation |
collection |
DOAJ |
language |
English |
format |
Article |
sources |
DOAJ |
author |
Malcolm M. Sande Mduduzi C. Hlophe Bodhaswar T. Maharaj |
spellingShingle |
Malcolm M. Sande Mduduzi C. Hlophe Bodhaswar T. Maharaj Access and Radio Resource Management for IAB Networks Using Deep Reinforcement Learning IEEE Access Congestion control deep reinforcement learning integrated access and backhaul millimeter wave nearest neighbor resource allocation |
author_facet |
Malcolm M. Sande Mduduzi C. Hlophe Bodhaswar T. Maharaj |
author_sort |
Malcolm M. Sande |
title |
Access and Radio Resource Management for IAB Networks Using Deep Reinforcement Learning |
title_short |
Access and Radio Resource Management for IAB Networks Using Deep Reinforcement Learning |
title_full |
Access and Radio Resource Management for IAB Networks Using Deep Reinforcement Learning |
title_fullStr |
Access and Radio Resource Management for IAB Networks Using Deep Reinforcement Learning |
title_full_unstemmed |
Access and Radio Resource Management for IAB Networks Using Deep Reinforcement Learning |
title_sort |
access and radio resource management for iab networks using deep reinforcement learning |
publisher |
IEEE |
series |
IEEE Access |
issn |
2169-3536 |
publishDate |
2021-01-01 |
description |
Congestion in dense traffic networks is a prominent obstacle towards realizing the performance requirements of 5G new radio. Since traditional adaptive traffic signal control cannot resolve this type of congestion, realizing context in the network and adapting resource allocation based on real-time parameters is an attractive approach. This article proposes a radio resource management solution for congestion avoidance on the access side of an integrated access and backhaul (IAB) network using deep reinforcement learning (DRL). The objective of this article is to obtain an optimal policy under which the transmission throughput of all UEs is maximized under the dictates of environmental pressures such as traffic load and transmission power. Here, the resource management problem was converted into a constrained problem using Markov decision processes and dynamic power management, where a deep neural network was trained for optimal power allocation. By initializing a power control parameter, <inline-formula> <tex-math notation="LaTeX">$\theta _{t}$ </tex-math></inline-formula>, with zero-mean normal distribution, the DRL algorithm adopts a learning policy that aims to achieve logical allocation of resources by placing more emphasis on congestion control and user satisfaction. The performance of the proposed DRL algorithm was evaluated using two learning schemes, i.e., individual learning and nearest neighbor cooperative learning, and this was compared with the performance of a baseline algorithm. The simulation results indicate that the proposed algorithms give better overall performance when compared to the baseline algorithm. From the simulation results, there is a subtle, but critically important insight that brings into focus the fundamental connection between learning rate and the two proposed algorithms. The nearest neighbor cooperative learning algorithm is suitable for IAB networks because its throughput has a good correlation with the congestion rate. |
topic |
Congestion control deep reinforcement learning integrated access and backhaul millimeter wave nearest neighbor resource allocation |
url |
https://ieeexplore.ieee.org/document/9512096/ |
work_keys_str_mv |
AT malcolmmsande accessandradioresourcemanagementforiabnetworksusingdeepreinforcementlearning AT mduduzichlophe accessandradioresourcemanagementforiabnetworksusingdeepreinforcementlearning AT bodhaswartmaharaj accessandradioresourcemanagementforiabnetworksusingdeepreinforcementlearning |
_version_ |
1721200781054443520 |