The Community for Technology Leaders
Parallel and Distributed Processing Symposium, International (2004)
Santa Fe, New Mexico
Apr. 26, 2004 to Apr. 30, 2004
ISBN: 0-7695-2132-0
pp: 249b
P. D. Coddington , University of Adelaide
D. A. Grove , University of Adelaide
This paper gives an overview of two related tools that we have developed to provide more accurate measurement and modelling of the performance of message passing programs and communications on distributed memory parallel computers. MPIBench uses a very precise, globally synchronised clock to measure the performance of MPI communication routines, and can generate probability distributions of communication times, not just the average values produced by other MPI benchmarks. This allows useful insights into MPI communications performance of parallel computers, particularly the effects of network contention. PEVPM provides a simple, fast and accurate technique for performance modelling and prediction of message-passing parallel programs. It uses a virtual parallel machine to simulate the execution of the parallel program. The effects of network contention can be accurately modelled by sampling from the probability distributions generated by MPIBench. These tools are particularly useful on Beowulf clusters with commodity Ethernet networks, where relatively high latencies, network congestion and TCP problems can significantly affect communication performance, and can be difficult to model accurately using other tools. Experiments with example parallel programs demonstrate that PEVPM gives accurate performance predictions on Beowulf clusters. We also show that modelling communication performance using average times rather than sampling from probability distributions can give misleading results, particularly for a large number of processors.
P. D. Coddington, D. A. Grove, "Communication Benchmarking and Performance Modelling of MPI Programs on Cluster Computers", Parallel and Distributed Processing Symposium, International, vol. 15, no. , pp. 249b, 2004, doi:10.1109/IPDPS.2004.1303309
104 ms
(Ver 3.3 (11022016))