2014 IEEE International Parallel & Distributed Processing Symposium Workshops (IPDPSW) (2014)
Phoenix, AZ, USA
May 19, 2014 to May 23, 2014
Krylov subspace solvers are often the method of choice when solving sparse linear systems iteratively. At the same time, hardware accelerators such as graphics processing units (GPUs) continue to offer significant floating point performance gains for matrix and vector computations through easy-to-use libraries of computational kernels. However, as these libraries are usually composed of a well optimized but limited set of linear algebra operations, applications that use them often fail to leverage the full potential of the accelerator. In this paper we target the acceleration of the BiCGSTAB solver for GPUs, showing that significant improvement can be achieved by reformulating the method and developing application-specific kernels instead of using the generic CUBLAS library provided by NVIDIA. We propose an implementation that benefits from a significantly reduced number of kernel launches and GPU-host communication events, by means of increased data locality and a simultaneous reduction of multiple scalar products. Using experimental data, we show that, depending on the dominance of the untouched sparse matrix vector products, significant performance improvements can be achieved compared to a reference implementation based on the CUBLAS library. We feel that such optimizations are crucial for the subsequent development of high-level sparse linear algebra libraries.
Vectors, Kernel, Graphics processing units, Sparse matrices, Libraries, Instruction sets, Linear systems
H. Anzt, W. Sawyer, S. Tomov, P. Luszczek, I. Yamazaki and J. Dongarra, "Optimizing Krylov Subspace Solvers on Graphics Processing Units," 2014 IEEE International Parallel & Distributed Processing Symposium Workshops (IPDPSW), Phoenix, AZ, USA, 2014, pp. 941-949.