QR-SDN: Towards Reinforcement Learning States, Actions, and Rewards for Direct Flow Routing in Software-Defined Networks
Flow routing can achieve fine-grained network performance optimizations by routing distinct packet traffic flows over different network paths. While the centralized control of Software-Defined Networking (SDN) provides a control framework for implementing centralized network optimizations, e.g., opt...
Main Authors: | , , , , |
---|---|
Format: | Article |
Language: | English |
Published: |
IEEE
2020-01-01
|
Series: | IEEE Access |
Subjects: | |
Online Access: | https://ieeexplore.ieee.org/document/9201294/ |
id |
doaj-6e25f30289dc4ab3bee1d3a854a83f36 |
---|---|
record_format |
Article |
spelling |
doaj-6e25f30289dc4ab3bee1d3a854a83f362021-03-30T04:25:34ZengIEEEIEEE Access2169-35362020-01-01817477317479110.1109/ACCESS.2020.30254329201294QR-SDN: Towards Reinforcement Learning States, Actions, and Rewards for Direct Flow Routing in Software-Defined NetworksJustus Rischke0https://orcid.org/0000-0001-9247-8156Peter Sossalla1https://orcid.org/0000-0003-1605-7581Hani Salah2https://orcid.org/0000-0002-1032-6659Frank H. P. Fitzek3https://orcid.org/0000-0001-8469-9573Martin Reisslein4https://orcid.org/0000-0003-1606-233XDeutsche Telekom Chair, 5G Lab Germany, Technische Universität Dresden, Dresden, GermanyDeutsche Telekom Chair, 5G Lab Germany, Technische Universität Dresden, Dresden, GermanyDeutsche Telekom Chair, 5G Lab Germany, Technische Universität Dresden, Dresden, GermanyDeutsche Telekom Chair, 5G Lab Germany, Technische Universität Dresden, Dresden, GermanySchool of Electrical, Computer, and Energy Engineering, Arizona State University, Tempe, AZ, USAFlow routing can achieve fine-grained network performance optimizations by routing distinct packet traffic flows over different network paths. While the centralized control of Software-Defined Networking (SDN) provides a control framework for implementing centralized network optimizations, e.g., optimized flow routing, the implementation of flow routing that is adaptive to varying traffic loads requires complex models. The goal of this study is to pursue a model-free approach that is based on reinforcement learning. We design and evaluate QR-SDN, a classical tabular reinforcement learning approach that directly represents the routing paths of individual flows in its state-action space. Due to the direct representation of flow routes in the QR-SDN state-action space, QR-SDN is the first reinforcement learning SDN routing approach to enable multiple routing paths between a given source (ingress) switch-destination (egress) switch pair while preserving the flow integrity. That is, in QR-SDN, packets of a given flow take the same routing path, while different flows with the same source-destination switch pair may take different routes (in contrast, the recent DRL-TE approach splits a given flow on a per-packet basis incurring high complexity and out-of-order packets). We implemented QR-SDN in a Software-Defined Network (SDN) emulation testbed. Our evaluations demonstrate that the flow-preserving multi-path routing of QR-SDN achieves substantially lower flow latencies than prior routing approaches that determine only a single source-destination route. A limitation of QR-SDN is that the state-action space grows exponentially with the number of network nodes. Addressing the scalability of direct flow routing, e.g., through routing only high-rate flows, is an important direction for future research. The QR-SDN code is made publicly available to support this future research.https://ieeexplore.ieee.org/document/9201294/Flow routingQ-tablesoftware-defined networking (SDN)state-space design |
collection |
DOAJ |
language |
English |
format |
Article |
sources |
DOAJ |
author |
Justus Rischke Peter Sossalla Hani Salah Frank H. P. Fitzek Martin Reisslein |
spellingShingle |
Justus Rischke Peter Sossalla Hani Salah Frank H. P. Fitzek Martin Reisslein QR-SDN: Towards Reinforcement Learning States, Actions, and Rewards for Direct Flow Routing in Software-Defined Networks IEEE Access Flow routing Q-table software-defined networking (SDN) state-space design |
author_facet |
Justus Rischke Peter Sossalla Hani Salah Frank H. P. Fitzek Martin Reisslein |
author_sort |
Justus Rischke |
title |
QR-SDN: Towards Reinforcement Learning States, Actions, and Rewards for Direct Flow Routing in Software-Defined Networks |
title_short |
QR-SDN: Towards Reinforcement Learning States, Actions, and Rewards for Direct Flow Routing in Software-Defined Networks |
title_full |
QR-SDN: Towards Reinforcement Learning States, Actions, and Rewards for Direct Flow Routing in Software-Defined Networks |
title_fullStr |
QR-SDN: Towards Reinforcement Learning States, Actions, and Rewards for Direct Flow Routing in Software-Defined Networks |
title_full_unstemmed |
QR-SDN: Towards Reinforcement Learning States, Actions, and Rewards for Direct Flow Routing in Software-Defined Networks |
title_sort |
qr-sdn: towards reinforcement learning states, actions, and rewards for direct flow routing in software-defined networks |
publisher |
IEEE |
series |
IEEE Access |
issn |
2169-3536 |
publishDate |
2020-01-01 |
description |
Flow routing can achieve fine-grained network performance optimizations by routing distinct packet traffic flows over different network paths. While the centralized control of Software-Defined Networking (SDN) provides a control framework for implementing centralized network optimizations, e.g., optimized flow routing, the implementation of flow routing that is adaptive to varying traffic loads requires complex models. The goal of this study is to pursue a model-free approach that is based on reinforcement learning. We design and evaluate QR-SDN, a classical tabular reinforcement learning approach that directly represents the routing paths of individual flows in its state-action space. Due to the direct representation of flow routes in the QR-SDN state-action space, QR-SDN is the first reinforcement learning SDN routing approach to enable multiple routing paths between a given source (ingress) switch-destination (egress) switch pair while preserving the flow integrity. That is, in QR-SDN, packets of a given flow take the same routing path, while different flows with the same source-destination switch pair may take different routes (in contrast, the recent DRL-TE approach splits a given flow on a per-packet basis incurring high complexity and out-of-order packets). We implemented QR-SDN in a Software-Defined Network (SDN) emulation testbed. Our evaluations demonstrate that the flow-preserving multi-path routing of QR-SDN achieves substantially lower flow latencies than prior routing approaches that determine only a single source-destination route. A limitation of QR-SDN is that the state-action space grows exponentially with the number of network nodes. Addressing the scalability of direct flow routing, e.g., through routing only high-rate flows, is an important direction for future research. The QR-SDN code is made publicly available to support this future research. |
topic |
Flow routing Q-table software-defined networking (SDN) state-space design |
url |
https://ieeexplore.ieee.org/document/9201294/ |
work_keys_str_mv |
AT justusrischke qrsdntowardsreinforcementlearningstatesactionsandrewardsfordirectflowroutinginsoftwaredefinednetworks AT petersossalla qrsdntowardsreinforcementlearningstatesactionsandrewardsfordirectflowroutinginsoftwaredefinednetworks AT hanisalah qrsdntowardsreinforcementlearningstatesactionsandrewardsfordirectflowroutinginsoftwaredefinednetworks AT frankhpfitzek qrsdntowardsreinforcementlearningstatesactionsandrewardsfordirectflowroutinginsoftwaredefinednetworks AT martinreisslein qrsdntowardsreinforcementlearningstatesactionsandrewardsfordirectflowroutinginsoftwaredefinednetworks |
_version_ |
1724181881364152320 |