Optimization for sparse matrix-vector multiplication based on NVIDIA CUDA platform

碩士 === 國立彰化師範大學 === 資訊工程學系 === 105 === In recent years, large size sparse matrices are often used in fields such as science and engineering which usually apply in computing linear model. Using the ELLPACK format to store sparse matrices, it can reduce the matrix storage space. But if there is too mu...

Full description

Bibliographic Details
Main Authors: Tsai,Sung-Han, 蔡松翰
Other Authors: Wei, Kai-Cheng
Format: Others
Language:zh-TW
Published: 2017
Online Access:http://ndltd.ncl.edu.tw/handle/qw23p7
id ndltd-TW-105NCUE5392016
record_format oai_dc
spelling ndltd-TW-105NCUE53920162019-05-16T00:15:44Z http://ndltd.ncl.edu.tw/handle/qw23p7 Optimization for sparse matrix-vector multiplication based on NVIDIA CUDA platform 以NVIDIA CUDA架構為基礎之稀疏矩陣乘積計算最佳化 Tsai,Sung-Han 蔡松翰 碩士 國立彰化師範大學 資訊工程學系 105 In recent years, large size sparse matrices are often used in fields such as science and engineering which usually apply in computing linear model. Using the ELLPACK format to store sparse matrices, it can reduce the matrix storage space. But if there is too much nonzero elements in one of row of the original sparse matrix, it still waste too much memory space. There are many research focusing on the Sparse Matrix–Vector Multiplication(SpMV)with ELLPACK format on Graphic Processing Unit(GPU). Therefore, the purpose of our research is reducing the access space of sparse matrix which is transformed in Compressed Sparse Row(CSR)format after Reverse Cutthill-McKee(RCM)algorithm to accelerate for SpMV on GPU. Due to lower data access ratio from SpMV, the performance is restricted by memory bandwidth. Our propose is based on CSR format from two aspects:(1)reduce cache misses to enhance the vector locality and raise the performance, and(2)reduce accessed matrix data by index reduction to optimize the performance. Wei, Kai-Cheng 魏凱城 2017 學位論文 ; thesis 34 zh-TW
collection NDLTD
language zh-TW
format Others
sources NDLTD
description 碩士 === 國立彰化師範大學 === 資訊工程學系 === 105 === In recent years, large size sparse matrices are often used in fields such as science and engineering which usually apply in computing linear model. Using the ELLPACK format to store sparse matrices, it can reduce the matrix storage space. But if there is too much nonzero elements in one of row of the original sparse matrix, it still waste too much memory space. There are many research focusing on the Sparse Matrix–Vector Multiplication(SpMV)with ELLPACK format on Graphic Processing Unit(GPU). Therefore, the purpose of our research is reducing the access space of sparse matrix which is transformed in Compressed Sparse Row(CSR)format after Reverse Cutthill-McKee(RCM)algorithm to accelerate for SpMV on GPU. Due to lower data access ratio from SpMV, the performance is restricted by memory bandwidth. Our propose is based on CSR format from two aspects:(1)reduce cache misses to enhance the vector locality and raise the performance, and(2)reduce accessed matrix data by index reduction to optimize the performance.
author2 Wei, Kai-Cheng
author_facet Wei, Kai-Cheng
Tsai,Sung-Han
蔡松翰
author Tsai,Sung-Han
蔡松翰
spellingShingle Tsai,Sung-Han
蔡松翰
Optimization for sparse matrix-vector multiplication based on NVIDIA CUDA platform
author_sort Tsai,Sung-Han
title Optimization for sparse matrix-vector multiplication based on NVIDIA CUDA platform
title_short Optimization for sparse matrix-vector multiplication based on NVIDIA CUDA platform
title_full Optimization for sparse matrix-vector multiplication based on NVIDIA CUDA platform
title_fullStr Optimization for sparse matrix-vector multiplication based on NVIDIA CUDA platform
title_full_unstemmed Optimization for sparse matrix-vector multiplication based on NVIDIA CUDA platform
title_sort optimization for sparse matrix-vector multiplication based on nvidia cuda platform
publishDate 2017
url http://ndltd.ncl.edu.tw/handle/qw23p7
work_keys_str_mv AT tsaisunghan optimizationforsparsematrixvectormultiplicationbasedonnvidiacudaplatform
AT càisōnghàn optimizationforsparsematrixvectormultiplicationbasedonnvidiacudaplatform
AT tsaisunghan yǐnvidiacudajiàgòuwèijīchǔzhīxīshūjǔzhènchéngjījìsuànzuìjiāhuà
AT càisōnghàn yǐnvidiacudajiàgòuwèijīchǔzhīxīshūjǔzhènchéngjījìsuànzuìjiāhuà
_version_ 1719163517975658496