The Community for Technology Leaders
Green Image
<p><b>Abstract</b>—We model a deterministic parallel program by a directed acyclic graph of tasks, where a task can execute as soon as all tasks preceding it have been executed. Each task can allocate or release an arbitrary amount of memory (i.e., heap memory allocation can be modeled). We call a parallel schedule "space efficient" if the amount of memory required is at most equal to the number of processors times the amount of memory required for some depth-first execution of the program by a single processor. We will describe a simple, locally depth-first, scheduling algorithm and show that it is always space efficient. Since the scheduling algorithm is greedy, it will be within a factor of two of being optimal with respect to time. For the special case of a program having a series-parallel structure, we show how to efficiently compute the worst case memory requirements over all possible depth-first executions of a program. Finally, we show how scheduling can be decentralized, making the approach scalable to a large number of processors when there is sufficient parallelism.</p>
Memory management, scheduling, worst case performance, parallel programming, memory bounds, shared memory.

D. J. Simpson and F. W. Burton, "Space Efficient Execution of Deterministic Parallel Programs," in IEEE Transactions on Software Engineering, vol. 25, no. , pp. 870-882, 1999.
101 ms
(Ver 3.3 (11022016))