Knowledge Transfer for On-Device Deep Reinforcement Learning in Resource Constrained Edge Computing Systems

Deep reinforcement learning (DRL) is a promising approach for developing control policies by learning how to perform tasks. Edge devices are required to control their actions by exploiting DRL to solve tasks autonomously in various applications such as smart manufacturing and autonomous driving. How...

Full description

Bibliographic Details
Main Authors: Ingook Jang, Hyunseok Kim, Donghun Lee, Young-Sung Son, Seonghyun Kim
Format: Article
Language:English
Published: IEEE 2020-01-01
Series:IEEE Access
Subjects:
Online Access:https://ieeexplore.ieee.org/document/9162017/
id doaj-5613f6ff2a2946caa76f8406e029b013
record_format Article
spelling doaj-5613f6ff2a2946caa76f8406e029b0132021-03-30T04:13:20ZengIEEEIEEE Access2169-35362020-01-01814658814659710.1109/ACCESS.2020.30149229162017Knowledge Transfer for On-Device Deep Reinforcement Learning in Resource Constrained Edge Computing SystemsIngook Jang0https://orcid.org/0000-0002-4804-6417Hyunseok Kim1Donghun Lee2Young-Sung Son3Seonghyun Kim4Autonomous IoT Research Section, Electronics and Telecommunications Research Institute (ETRI), Daejeon, South KoreaAutonomous IoT Research Section, Electronics and Telecommunications Research Institute (ETRI), Daejeon, South KoreaAutonomous IoT Research Section, Electronics and Telecommunications Research Institute (ETRI), Daejeon, South KoreaAutonomous IoT Research Section, Electronics and Telecommunications Research Institute (ETRI), Daejeon, South KoreaAutonomous IoT Research Section, Electronics and Telecommunications Research Institute (ETRI), Daejeon, South KoreaDeep reinforcement learning (DRL) is a promising approach for developing control policies by learning how to perform tasks. Edge devices are required to control their actions by exploiting DRL to solve tasks autonomously in various applications such as smart manufacturing and autonomous driving. However, the resource limitations of edge devices make it unfeasible for them to train their policies from scratch. It is also impractical for such an edge device to use the policy with a large number of layers and parameters, which is pre-trained by a centralized cloud infrastructure with high computational power. In this paper, we propose a method, on-device DRL with distillation (OD3), to efficiently transfer distilled knowledge of how to behave for on-device DRL in resource-constrained edge computing systems. Our proposed method makes it possible to simultaneously perform knowledge transfer and policy model compression in a single training process on edge devices with considering their limited resource budgets. The novelty of our method is to apply a knowledge distillation approach to DRL based edge device control in integrated edge cloud environments. We analyze the performance of the proposed method by implementing it on a commercial embedded system-on-module equipped with limited hardware resources. The experimental results show that 1) edge policy training with the proposed method achieves near-cloud-performance in terms of average rewards, although the size of the edge policy network is significantly smaller compared to that of the cloud policy network and 2) the training time elapsed for edge policy training with our method is reduced significantly compared to edge policy training from scratch.https://ieeexplore.ieee.org/document/9162017/Deep reinforcement learningedge computingedge AIknowledge transferpolicy model compressionon-device training
collection DOAJ
language English
format Article
sources DOAJ
author Ingook Jang
Hyunseok Kim
Donghun Lee
Young-Sung Son
Seonghyun Kim
spellingShingle Ingook Jang
Hyunseok Kim
Donghun Lee
Young-Sung Son
Seonghyun Kim
Knowledge Transfer for On-Device Deep Reinforcement Learning in Resource Constrained Edge Computing Systems
IEEE Access
Deep reinforcement learning
edge computing
edge AI
knowledge transfer
policy model compression
on-device training
author_facet Ingook Jang
Hyunseok Kim
Donghun Lee
Young-Sung Son
Seonghyun Kim
author_sort Ingook Jang
title Knowledge Transfer for On-Device Deep Reinforcement Learning in Resource Constrained Edge Computing Systems
title_short Knowledge Transfer for On-Device Deep Reinforcement Learning in Resource Constrained Edge Computing Systems
title_full Knowledge Transfer for On-Device Deep Reinforcement Learning in Resource Constrained Edge Computing Systems
title_fullStr Knowledge Transfer for On-Device Deep Reinforcement Learning in Resource Constrained Edge Computing Systems
title_full_unstemmed Knowledge Transfer for On-Device Deep Reinforcement Learning in Resource Constrained Edge Computing Systems
title_sort knowledge transfer for on-device deep reinforcement learning in resource constrained edge computing systems
publisher IEEE
series IEEE Access
issn 2169-3536
publishDate 2020-01-01
description Deep reinforcement learning (DRL) is a promising approach for developing control policies by learning how to perform tasks. Edge devices are required to control their actions by exploiting DRL to solve tasks autonomously in various applications such as smart manufacturing and autonomous driving. However, the resource limitations of edge devices make it unfeasible for them to train their policies from scratch. It is also impractical for such an edge device to use the policy with a large number of layers and parameters, which is pre-trained by a centralized cloud infrastructure with high computational power. In this paper, we propose a method, on-device DRL with distillation (OD3), to efficiently transfer distilled knowledge of how to behave for on-device DRL in resource-constrained edge computing systems. Our proposed method makes it possible to simultaneously perform knowledge transfer and policy model compression in a single training process on edge devices with considering their limited resource budgets. The novelty of our method is to apply a knowledge distillation approach to DRL based edge device control in integrated edge cloud environments. We analyze the performance of the proposed method by implementing it on a commercial embedded system-on-module equipped with limited hardware resources. The experimental results show that 1) edge policy training with the proposed method achieves near-cloud-performance in terms of average rewards, although the size of the edge policy network is significantly smaller compared to that of the cloud policy network and 2) the training time elapsed for edge policy training with our method is reduced significantly compared to edge policy training from scratch.
topic Deep reinforcement learning
edge computing
edge AI
knowledge transfer
policy model compression
on-device training
url https://ieeexplore.ieee.org/document/9162017/
work_keys_str_mv AT ingookjang knowledgetransferforondevicedeepreinforcementlearninginresourceconstrainededgecomputingsystems
AT hyunseokkim knowledgetransferforondevicedeepreinforcementlearninginresourceconstrainededgecomputingsystems
AT donghunlee knowledgetransferforondevicedeepreinforcementlearninginresourceconstrainededgecomputingsystems
AT youngsungson knowledgetransferforondevicedeepreinforcementlearninginresourceconstrainededgecomputingsystems
AT seonghyunkim knowledgetransferforondevicedeepreinforcementlearninginresourceconstrainededgecomputingsystems
_version_ 1724182192435757056