Abstract | ||
---|---|---|
Irregular computations on large workloads are a necessity in many areas of computational science. Mapping these computations to modern parallel architectures, such as GPUs, is particularly challenging because the performance often depends critically on the choice of data-structure and algorithm. In this paper, we develop a parallel processing scheme, based on Merge Path partitioning, to compute segmented row-wise operations on sparse matrices that exposes parallelism at the granularity of individual nonzero entries. Our decomposition achieves competitive performance across many diverse problems while maintaining predictable behaviour dependent only on the computational work and ameliorates the impact of irregularity. We evaluate the performance of three sparse kernels: Spiv, Spaded and Sperm. We show that our processing scheme for each kernel yields comparable performance to other schemes in many cases and our performance is highly correlated, nearly 1, to the computational work irrespective of the underlying structure of the matrices. |
Year | DOI | Venue |
---|---|---|
2015 | 10.1109/IPDPS.2015.98 | International Parallel & Distributed Processing Symposium |
Keywords | Field | DocType |
parallel, gpu, sparse matrix-matrix, sorting | Kernel (linear algebra),Matrix (mathematics),Computer science,Parallel computing,Sparse approximation,Sorting,Granularity,Sparse matrix,Matrix-free methods,Distributed computing,Computation | Conference |
ISSN | Citations | PageRank |
1530-2075 | 4 | 0.40 |
References | Authors | |
11 | 5 |
Name | Order | Citations | PageRank |
---|---|---|---|
Steven Dalton | 1 | 90 | 3.99 |
Sean Baxter | 2 | 13 | 1.65 |
Duane Merrill | 3 | 283 | 11.82 |
Luke Olson | 4 | 235 | 21.93 |
Michael Garland | 5 | 4328 | 239.08 |