The Community for Technology Leaders
Parallel and Distributed Processing Symposium, International (2013)
Cambridge, MA, USA USA
May 20, 2013 to May 24, 2013
ISSN: 1530-2075
ISBN: 978-1-4673-6066-1
pp: 560-568
Reductions matter and they are here to stay. Wide adoption of parallel processing hardware in a broad range of computer applications has encouraged recent research efforts on their efficient parallelization. Furthermore, trends towards high productivity languages in mainstream computing increases the demand for efficient programming support. In this paper we present a new approach on parallel reductions for distributed memory systems that provides both scalability and programmability. Using OmpSs, a task-based parallel programming model, the developer has the ability to express scalable reductions through a single pragma annotation. This pragma annotation is applicable for tasks as well as for work-sharing constructs (with implicit tasking) and instructs the compiler to generate the required runtime calls. The supporting runtime handles data and task distribution, parallel execution and data reduction. Scalability is achieved through a software cache that maximizes local and temporal data reuse and allows overlapped computation and communication. Results confirm scalability for up to 32 12-core cluster nodes.
Runtime, Histograms, Vectors, Software, Reactive power, Arrays, Scalability, distributed systems, reductions, software cache, parallel programming, runtime systems

J. Ciesko, J. Bueno, N. Puzovic, A. Ramirez, R. M. Badia and J. Labarta, "Programmable and Scalable Reductions on Clusters," 2013 IEEE 27th International Symposium on Parallel and Distributed Processing (IPDPS 2013)(IPDPS), Boston, MA, 2013, pp. 560-568.
88 ms
(Ver 3.3 (11022016))