A Two-Stage Task Scheduling Scheme on CUDA Systems
碩士 === 國立成功大學 === 資訊工程學系碩博士班 === 98 === Recently graphics processing units (GPUs) have become an important high-performance computing platform. In many application fields, using GPU to accelerate computation has proven to be feasible. However, many independent tasks do not fully utilize the GPU reso...
Main Authors: | , |
---|---|
Other Authors: | |
Format: | Others |
Language: | zh-TW |
Published: |
2010
|
Online Access: | http://ndltd.ncl.edu.tw/handle/44501772872689956791 |
id |
ndltd-TW-098NCKU5392069 |
---|---|
record_format |
oai_dc |
spelling |
ndltd-TW-098NCKU53920692015-11-06T04:04:00Z http://ndltd.ncl.edu.tw/handle/44501772872689956791 A Two-Stage Task Scheduling Scheme on CUDA Systems 在CUDA系統上兩階段任務排程方法 Yu-ChuHuang 黃于鑄 碩士 國立成功大學 資訊工程學系碩博士班 98 Recently graphics processing units (GPUs) have become an important high-performance computing platform. In many application fields, using GPU to accelerate computation has proven to be feasible. However, many independent tasks do not fully utilize the GPU resources, suggesting scheduling independent tasks is an important issue worth to be studied for platform with GPUs. This thesis proposes a two-stage task scheduling scheme on CUDA systems. In the first stage, the task allocation problem is mapped to the bin packing problem and the First-Fit Decreasing algorithm is selected to solve it. In the second stage, we use the asynchronous transfers and overlap transfers with computation to execute tasks. Base on the experimental results, we show our method is on average 1.57 times faster than the sequential method and 1.2 times faster than the merge kernel method proposed [9] by Guevara et al. Chih-Ping Chu 朱治平 2010 學位論文 ; thesis 46 zh-TW |
collection |
NDLTD |
language |
zh-TW |
format |
Others
|
sources |
NDLTD |
description |
碩士 === 國立成功大學 === 資訊工程學系碩博士班 === 98 === Recently graphics processing units (GPUs) have become an important high-performance computing platform. In many application fields, using GPU to accelerate computation has proven to be feasible. However, many independent tasks do not fully utilize the GPU resources, suggesting scheduling independent tasks is an important issue worth to be studied for platform with GPUs. This thesis proposes a two-stage task scheduling scheme on CUDA systems. In the first stage, the task allocation problem is mapped to the bin packing problem and the First-Fit Decreasing algorithm is selected to solve it. In the second stage, we use the asynchronous transfers and overlap transfers with computation to execute tasks. Base on the experimental results, we show our method is on average 1.57 times faster than the sequential method and 1.2 times faster than the merge kernel method proposed [9] by Guevara et al.
|
author2 |
Chih-Ping Chu |
author_facet |
Chih-Ping Chu Yu-ChuHuang 黃于鑄 |
author |
Yu-ChuHuang 黃于鑄 |
spellingShingle |
Yu-ChuHuang 黃于鑄 A Two-Stage Task Scheduling Scheme on CUDA Systems |
author_sort |
Yu-ChuHuang |
title |
A Two-Stage Task Scheduling Scheme on CUDA Systems |
title_short |
A Two-Stage Task Scheduling Scheme on CUDA Systems |
title_full |
A Two-Stage Task Scheduling Scheme on CUDA Systems |
title_fullStr |
A Two-Stage Task Scheduling Scheme on CUDA Systems |
title_full_unstemmed |
A Two-Stage Task Scheduling Scheme on CUDA Systems |
title_sort |
two-stage task scheduling scheme on cuda systems |
publishDate |
2010 |
url |
http://ndltd.ncl.edu.tw/handle/44501772872689956791 |
work_keys_str_mv |
AT yuchuhuang atwostagetaskschedulingschemeoncudasystems AT huángyúzhù atwostagetaskschedulingschemeoncudasystems AT yuchuhuang zàicudaxìtǒngshàngliǎngjiēduànrènwùpáichéngfāngfǎ AT huángyúzhù zàicudaxìtǒngshàngliǎngjiēduànrènwùpáichéngfāngfǎ AT yuchuhuang twostagetaskschedulingschemeoncudasystems AT huángyúzhù twostagetaskschedulingschemeoncudasystems |
_version_ |
1718125917886218240 |