A Two-Stage Task Scheduling Scheme on CUDA Systems

碩士 === 國立成功大學 === 資訊工程學系碩博士班 === 98 === Recently graphics processing units (GPUs) have become an important high-performance computing platform. In many application fields, using GPU to accelerate computation has proven to be feasible. However, many independent tasks do not fully utilize the GPU reso...

Full description

Bibliographic Details
Main Authors: Yu-ChuHuang, 黃于鑄
Other Authors: Chih-Ping Chu
Format: Others
Language:zh-TW
Published: 2010
Online Access:http://ndltd.ncl.edu.tw/handle/44501772872689956791
id ndltd-TW-098NCKU5392069
record_format oai_dc
spelling ndltd-TW-098NCKU53920692015-11-06T04:04:00Z http://ndltd.ncl.edu.tw/handle/44501772872689956791 A Two-Stage Task Scheduling Scheme on CUDA Systems 在CUDA系統上兩階段任務排程方法 Yu-ChuHuang 黃于鑄 碩士 國立成功大學 資訊工程學系碩博士班 98 Recently graphics processing units (GPUs) have become an important high-performance computing platform. In many application fields, using GPU to accelerate computation has proven to be feasible. However, many independent tasks do not fully utilize the GPU resources, suggesting scheduling independent tasks is an important issue worth to be studied for platform with GPUs. This thesis proposes a two-stage task scheduling scheme on CUDA systems. In the first stage, the task allocation problem is mapped to the bin packing problem and the First-Fit Decreasing algorithm is selected to solve it. In the second stage, we use the asynchronous transfers and overlap transfers with computation to execute tasks. Base on the experimental results, we show our method is on average 1.57 times faster than the sequential method and 1.2 times faster than the merge kernel method proposed [9] by Guevara et al. Chih-Ping Chu 朱治平 2010 學位論文 ; thesis 46 zh-TW
collection NDLTD
language zh-TW
format Others
sources NDLTD
description 碩士 === 國立成功大學 === 資訊工程學系碩博士班 === 98 === Recently graphics processing units (GPUs) have become an important high-performance computing platform. In many application fields, using GPU to accelerate computation has proven to be feasible. However, many independent tasks do not fully utilize the GPU resources, suggesting scheduling independent tasks is an important issue worth to be studied for platform with GPUs. This thesis proposes a two-stage task scheduling scheme on CUDA systems. In the first stage, the task allocation problem is mapped to the bin packing problem and the First-Fit Decreasing algorithm is selected to solve it. In the second stage, we use the asynchronous transfers and overlap transfers with computation to execute tasks. Base on the experimental results, we show our method is on average 1.57 times faster than the sequential method and 1.2 times faster than the merge kernel method proposed [9] by Guevara et al.
author2 Chih-Ping Chu
author_facet Chih-Ping Chu
Yu-ChuHuang
黃于鑄
author Yu-ChuHuang
黃于鑄
spellingShingle Yu-ChuHuang
黃于鑄
A Two-Stage Task Scheduling Scheme on CUDA Systems
author_sort Yu-ChuHuang
title A Two-Stage Task Scheduling Scheme on CUDA Systems
title_short A Two-Stage Task Scheduling Scheme on CUDA Systems
title_full A Two-Stage Task Scheduling Scheme on CUDA Systems
title_fullStr A Two-Stage Task Scheduling Scheme on CUDA Systems
title_full_unstemmed A Two-Stage Task Scheduling Scheme on CUDA Systems
title_sort two-stage task scheduling scheme on cuda systems
publishDate 2010
url http://ndltd.ncl.edu.tw/handle/44501772872689956791
work_keys_str_mv AT yuchuhuang atwostagetaskschedulingschemeoncudasystems
AT huángyúzhù atwostagetaskschedulingschemeoncudasystems
AT yuchuhuang zàicudaxìtǒngshàngliǎngjiēduànrènwùpáichéngfāngfǎ
AT huángyúzhù zàicudaxìtǒngshàngliǎngjiēduànrènwùpáichéngfāngfǎ
AT yuchuhuang twostagetaskschedulingschemeoncudasystems
AT huángyúzhù twostagetaskschedulingschemeoncudasystems
_version_ 1718125917886218240