Access and Radio Resource Management for IAB Networks Using Deep Reinforcement Learning

Congestion in dense traffic networks is a prominent obstacle towards realizing the performance requirements of 5G new radio. Since traditional adaptive traffic signal control cannot resolve this type of congestion, realizing context in the network and adapting resource allocation based on real-time...

Full description

Bibliographic Details
Main Authors: Malcolm M. Sande, Mduduzi C. Hlophe, Bodhaswar T. Maharaj
Format: Article
Language:English
Published: IEEE 2021-01-01
Series:IEEE Access
Subjects:
Online Access:https://ieeexplore.ieee.org/document/9512096/
id doaj-c3e07ecd2cad43b098dd2258f5a1fd13
record_format Article
spelling doaj-c3e07ecd2cad43b098dd2258f5a1fd132021-08-20T23:00:23ZengIEEEIEEE Access2169-35362021-01-01911421811423410.1109/ACCESS.2021.31043229512096Access and Radio Resource Management for IAB Networks Using Deep Reinforcement LearningMalcolm M. Sande0https://orcid.org/0000-0002-6636-2050Mduduzi C. Hlophe1https://orcid.org/0000-0001-6111-5619Bodhaswar T. Maharaj2https://orcid.org/0000-0002-3703-3637Department of Electrical, Electronic and Computer Engineering, University of Pretoria, Pretoria, South AfricaDepartment of Electrical, Electronic and Computer Engineering, University of Pretoria, Pretoria, South AfricaDepartment of Electrical, Electronic and Computer Engineering, University of Pretoria, Pretoria, South AfricaCongestion in dense traffic networks is a prominent obstacle towards realizing the performance requirements of 5G new radio. Since traditional adaptive traffic signal control cannot resolve this type of congestion, realizing context in the network and adapting resource allocation based on real-time parameters is an attractive approach. This article proposes a radio resource management solution for congestion avoidance on the access side of an integrated access and backhaul (IAB) network using deep reinforcement learning (DRL). The objective of this article is to obtain an optimal policy under which the transmission throughput of all UEs is maximized under the dictates of environmental pressures such as traffic load and transmission power. Here, the resource management problem was converted into a constrained problem using Markov decision processes and dynamic power management, where a deep neural network was trained for optimal power allocation. By initializing a power control parameter, <inline-formula> <tex-math notation="LaTeX">$\theta _{t}$ </tex-math></inline-formula>, with zero-mean normal distribution, the DRL algorithm adopts a learning policy that aims to achieve logical allocation of resources by placing more emphasis on congestion control and user satisfaction. The performance of the proposed DRL algorithm was evaluated using two learning schemes, i.e., individual learning and nearest neighbor cooperative learning, and this was compared with the performance of a baseline algorithm. The simulation results indicate that the proposed algorithms give better overall performance when compared to the baseline algorithm. From the simulation results, there is a subtle, but critically important insight that brings into focus the fundamental connection between learning rate and the two proposed algorithms. The nearest neighbor cooperative learning algorithm is suitable for IAB networks because its throughput has a good correlation with the congestion rate.https://ieeexplore.ieee.org/document/9512096/Congestion controldeep reinforcement learningintegrated access and backhaulmillimeter wavenearest neighborresource allocation
collection DOAJ
language English
format Article
sources DOAJ
author Malcolm M. Sande
Mduduzi C. Hlophe
Bodhaswar T. Maharaj
spellingShingle Malcolm M. Sande
Mduduzi C. Hlophe
Bodhaswar T. Maharaj
Access and Radio Resource Management for IAB Networks Using Deep Reinforcement Learning
IEEE Access
Congestion control
deep reinforcement learning
integrated access and backhaul
millimeter wave
nearest neighbor
resource allocation
author_facet Malcolm M. Sande
Mduduzi C. Hlophe
Bodhaswar T. Maharaj
author_sort Malcolm M. Sande
title Access and Radio Resource Management for IAB Networks Using Deep Reinforcement Learning
title_short Access and Radio Resource Management for IAB Networks Using Deep Reinforcement Learning
title_full Access and Radio Resource Management for IAB Networks Using Deep Reinforcement Learning
title_fullStr Access and Radio Resource Management for IAB Networks Using Deep Reinforcement Learning
title_full_unstemmed Access and Radio Resource Management for IAB Networks Using Deep Reinforcement Learning
title_sort access and radio resource management for iab networks using deep reinforcement learning
publisher IEEE
series IEEE Access
issn 2169-3536
publishDate 2021-01-01
description Congestion in dense traffic networks is a prominent obstacle towards realizing the performance requirements of 5G new radio. Since traditional adaptive traffic signal control cannot resolve this type of congestion, realizing context in the network and adapting resource allocation based on real-time parameters is an attractive approach. This article proposes a radio resource management solution for congestion avoidance on the access side of an integrated access and backhaul (IAB) network using deep reinforcement learning (DRL). The objective of this article is to obtain an optimal policy under which the transmission throughput of all UEs is maximized under the dictates of environmental pressures such as traffic load and transmission power. Here, the resource management problem was converted into a constrained problem using Markov decision processes and dynamic power management, where a deep neural network was trained for optimal power allocation. By initializing a power control parameter, <inline-formula> <tex-math notation="LaTeX">$\theta _{t}$ </tex-math></inline-formula>, with zero-mean normal distribution, the DRL algorithm adopts a learning policy that aims to achieve logical allocation of resources by placing more emphasis on congestion control and user satisfaction. The performance of the proposed DRL algorithm was evaluated using two learning schemes, i.e., individual learning and nearest neighbor cooperative learning, and this was compared with the performance of a baseline algorithm. The simulation results indicate that the proposed algorithms give better overall performance when compared to the baseline algorithm. From the simulation results, there is a subtle, but critically important insight that brings into focus the fundamental connection between learning rate and the two proposed algorithms. The nearest neighbor cooperative learning algorithm is suitable for IAB networks because its throughput has a good correlation with the congestion rate.
topic Congestion control
deep reinforcement learning
integrated access and backhaul
millimeter wave
nearest neighbor
resource allocation
url https://ieeexplore.ieee.org/document/9512096/
work_keys_str_mv AT malcolmmsande accessandradioresourcemanagementforiabnetworksusingdeepreinforcementlearning
AT mduduzichlophe accessandradioresourcemanagementforiabnetworksusingdeepreinforcementlearning
AT bodhaswartmaharaj accessandradioresourcemanagementforiabnetworksusingdeepreinforcementlearning
_version_ 1721200781054443520