Optimizing nonzero-based sparse matrix partitioning models via reducing latency

Limited Access
This item is unavailable until:
2020-12-01

Date

2018

Editor(s)

Advisor

Supervisor

Co-Advisor

Co-Supervisor

Instructor

Source Title

Journal of Parallel and Distributed Computing

Print ISSN

0743-7315

Electronic ISSN

Publisher

Academic Press

Volume

122

Issue

Pages

145 - 158

Language

English

Journal Title

Journal ISSN

Volume Title

Series

Abstract

For the parallelization of sparse matrix-vector multiplication (SpMV) on distributed memory systems, nonzero-based fine-grain and medium-grain partitioning models attain the lowest communication volume and computational imbalance among all partitioning models. This usually comes, however, at the expense of high message count, i.e., high latency overhead. This work addresses this shortcoming by proposing new fine-grain and medium-grain models that are able to minimize communication volume and message count in a single partitioning phase. The new models utilize message nets in order to encapsulate the minimization of total message count. We further fine-tune these models by proposing delayed addition and thresholding for message nets in order to establish a trade-off between the conflicting objectives of minimizing communication volume and message count. The experiments on an extensive dataset of nearly one thousand matrices show that the proposed models improve the total message count of the original nonzero-based models by up to 27% on the average, which is reflected on the parallel runtime of SpMV as an average reduction of 15% on 512 processors.

Course

Other identifiers

Book Title

Citation