The Development of High-Utilization Scheduling for Malleable Tasks Using Deep Reinforcement Learning
碩士 === 國立交通大學 === 資訊科學與工程研究所 === 107 === Modern high-performance computing platforms can perform dynamic tasks using elastic resource provisioning. However, complex schedules lead to useless resource fragments from time to time in the system. This paper employs our previous developed system to colle...
Main Authors: | , |
---|---|
Other Authors: | |
Format: | Others |
Language: | zh-TW |
Published: |
2019
|
Online Access: | http://ndltd.ncl.edu.tw/handle/pds4at |
id |
ndltd-TW-107NCTU5394138 |
---|---|
record_format |
oai_dc |
spelling |
ndltd-TW-107NCTU53941382019-11-26T05:16:56Z http://ndltd.ncl.edu.tw/handle/pds4at The Development of High-Utilization Scheduling for Malleable Tasks Using Deep Reinforcement Learning 運用強化學習發展高資源使用率之延展性工作排程方法 Chang, Yen-Ling 張晏菱 碩士 國立交通大學 資訊科學與工程研究所 107 Modern high-performance computing platforms can perform dynamic tasks using elastic resource provisioning. However, complex schedules lead to useless resource fragments from time to time in the system. This paper employs our previous developed system to collect resource fragments on the computing farm. The goal is to increasing resource utilization by using these resource fragments to perform lightweight malleable tasks. This paper investigates the efficient approach to assign a set of malleable tasks to a group of resource fragments. We propose a threshold calculation method. The threshold value is used to estimate the success rate for matching a task length to a type of resource fragment. The previous threshold algorithms were calculated using fixed formulas or statistical data and were not able to adapt to changing environments. In this paper, we adopt the PPO reinforcement learning method to train the correlation between system state and threshold values, and obtain better results than that of previous approaches. Wu, I-Chen Chang, Ming-Feng 吳毅成 張明峰 2019 學位論文 ; thesis 18 zh-TW |
collection |
NDLTD |
language |
zh-TW |
format |
Others
|
sources |
NDLTD |
description |
碩士 === 國立交通大學 === 資訊科學與工程研究所 === 107 === Modern high-performance computing platforms can perform dynamic tasks using elastic resource provisioning. However, complex schedules lead to useless resource fragments from time to time in the system. This paper employs our previous developed system to collect resource fragments on the computing farm. The goal is to increasing resource utilization by using these resource fragments to perform lightweight malleable tasks. This paper investigates the efficient approach to assign a set of malleable tasks to a group of resource fragments. We propose a threshold calculation method. The threshold value is used to estimate the success rate for matching a task length to a type of resource fragment. The previous threshold algorithms were calculated using fixed formulas or statistical data and were not able to adapt to changing environments. In this paper, we adopt the PPO reinforcement learning method to train the correlation between system state and threshold values, and obtain better results than that of previous approaches.
|
author2 |
Wu, I-Chen |
author_facet |
Wu, I-Chen Chang, Yen-Ling 張晏菱 |
author |
Chang, Yen-Ling 張晏菱 |
spellingShingle |
Chang, Yen-Ling 張晏菱 The Development of High-Utilization Scheduling for Malleable Tasks Using Deep Reinforcement Learning |
author_sort |
Chang, Yen-Ling |
title |
The Development of High-Utilization Scheduling for Malleable Tasks Using Deep Reinforcement Learning |
title_short |
The Development of High-Utilization Scheduling for Malleable Tasks Using Deep Reinforcement Learning |
title_full |
The Development of High-Utilization Scheduling for Malleable Tasks Using Deep Reinforcement Learning |
title_fullStr |
The Development of High-Utilization Scheduling for Malleable Tasks Using Deep Reinforcement Learning |
title_full_unstemmed |
The Development of High-Utilization Scheduling for Malleable Tasks Using Deep Reinforcement Learning |
title_sort |
development of high-utilization scheduling for malleable tasks using deep reinforcement learning |
publishDate |
2019 |
url |
http://ndltd.ncl.edu.tw/handle/pds4at |
work_keys_str_mv |
AT changyenling thedevelopmentofhighutilizationschedulingformalleabletasksusingdeepreinforcementlearning AT zhāngyànlíng thedevelopmentofhighutilizationschedulingformalleabletasksusingdeepreinforcementlearning AT changyenling yùnyòngqiánghuàxuéxífāzhǎngāozīyuánshǐyònglǜzhīyánzhǎnxìnggōngzuòpáichéngfāngfǎ AT zhāngyànlíng yùnyòngqiánghuàxuéxífāzhǎngāozīyuánshǐyònglǜzhīyánzhǎnxìnggōngzuòpáichéngfāngfǎ AT changyenling developmentofhighutilizationschedulingformalleabletasksusingdeepreinforcementlearning AT zhāngyànlíng developmentofhighutilizationschedulingformalleabletasksusingdeepreinforcementlearning |
_version_ |
1719296431311814656 |