DocumentCode :
131331
Title :
A memory transaction model for Sparse Matrix-Vector multiplications on GPUs
Author :
Keklikian, Thalie ; Pierre Langlois, J.M. ; Savaria, Yvon
Author_Institution :
Groupe de Rech. en Microelectron. et Microsystemes, Polytech. Montreal, Montréal, QC, Canada
fYear :
2014
fDate :
22-25 June 2014
Firstpage :
309
Lastpage :
312
Abstract :
The Sparse Matrix-Vector multiplication (SpMV) is an algorithm used in many fields. Since the introduction of CUDA and general purpose programming on GPUs, several efforts to optimize it have been reported. SpMV optimization is complex due to irregular memory accesses depending on the nonzero element distribution of the matrix. In this paper, we propose a model that predicts the number of memory transactions of SpMV for a matrix stored in the CSR format. With the number of memory transactions known in advance, the performance and the execution time can be estimated. The model can be used to select the best suited CUDA implementation for sparse matrices for a given application domain. Predicted results from the model are within 7.5% for the matrices of more than 1000 rows that we have tested on the NVIDIA Tesla K20c and Ge-Force GTX 670.
Keywords :
graphics processing units; matrix multiplication; storage management; vectors; CSR format; CUDA; GPU; NVIDIA Ge-Force GTX 670; NVIDIA Tesla K20c; SpMV algorithm; compute unified device architecture; general purpose programming; graphics processing unit; memory transaction model; nonzero element distribution; sparse matrix-vector multiplications; Graphics processing units; Instruction sets; Kernel; Optimization; Predictive models; Sparse matrices; Vectors; CUDA; GPU; SpMV;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
New Circuits and Systems Conference (NEWCAS), 2014 IEEE 12th International
Conference_Location :
Trois-Rivieres, QC
Type :
conf
DOI :
10.1109/NEWCAS.2014.6934044
Filename :
6934044
Link To Document :
بازگشت