Adaptive runtime exploiting sparsity in tensor of deep learning on heterogeneous systems

碩士 === 國立臺灣大學 === 資訊工程學研究所 === 105 === Heterogeneous computing achieves high performance by exploiting high parallelism and special type of computation (such as SIMD operations) available in applications on best fit computation devices. For example, massive and regular SIMD operations can be more ef...

Full description

Bibliographic Details
Main Authors: Kuo-You Peng, 彭國祐
Other Authors: Wei-Chung Hsu
Format: Others
Language:en_US
Published: 2017
Online Access:http://ndltd.ncl.edu.tw/handle/46739294222658497904
Description
Summary:碩士 === 國立臺灣大學 === 資訊工程學研究所 === 105 === Heterogeneous computing achieves high performance by exploiting high parallelism and special type of computation (such as SIMD operations) available in applications on best fit computation devices. For example, massive and regular SIMD operations can be more efficiently computed on GPU. However, the performance of heterogeneous program can be degraded when the portion assigned to GPU encounters irregular tasks. Deep learning is an application that has the characteristics of high parallelism but may also encounter irregular tasks. This study introduces a method which could reduce computation and improve the performance in the deep learning application by recording information on the runtime. By using collected information, we can adaptive changing workload when we encounter irregular tasks. When deep learning encounters irregular tasks, using our method could split the workload of the deep learning application into two parts: dense workload and sparse workload. The dense workload will be deployed on the GPU device, and the sparse part is sent for the CPU. In this way, GPU gets better computing efficiency, and CPU is more competent in handling sparse part than GPU.