Electronics | |
Performance Analysis of Sparse Matrix-Vector Multiplication (SpMV) on Graphics Processing Units (GPUs) | |
Thaha Mohammed1  Aiiad Albeshri2  Iyad Katib2  Sarah AlAhmadi3  Rashid Mehmood4  | |
[1] Department of Computer Science, Aalto University, 02150 Espoo, Finland;Department of Computer Science, King Abdulaziz University, Jeddah 21589, Saudi Arabia;Department of Computer and Information Sciences, Taibah University, Medina 42353, Saudi Arabia;High Performance Computing Center, King Abdulaziz University, Jeddah 21589, Saudi Arabia; | |
关键词: sparse matrix-vector multiplication (SpMV); high performance computing (HPC); sparse matrix storage; graphics processing units (GPUs); CSR; ELL; | |
DOI : 10.3390/electronics9101675 | |
来源: DOAJ |
【 摘 要 】
Graphics processing units (GPUs) have delivered a remarkable performance for a variety of high performance computing (HPC) applications through massive parallelism. One such application is sparse matrix-vector (SpMV) computations, which is central to many scientific, engineering, and other applications including machine learning. No single SpMV storage or computation scheme provides consistent and sufficiently high performance for all matrices due to their varying sparsity patterns. An extensive literature review reveals that the performance of SpMV techniques on GPUs has not been studied in sufficient detail. In this paper, we provide a detailed performance analysis of SpMV performance on GPUs using four notable sparse matrix storage schemes (compressed sparse row (CSR), ELLAPCK (ELL), hybrid ELL/COO (HYB), and compressed sparse row 5 (CSR5)), five performance metrics (execution time, giga floating point operations per second (GFLOPS), achieved occupancy, instructions per warp, and warp execution efficiency), five matrix sparsity features (
【 授权许可】
Unknown