
This Article  
 
Share  
Bibliographic References  
Add to:  
Digg Furl Spurl Blink Simpy Del.icio.us Y!MyWeb  
Search  
 
ASCII Text  x  
T. Yang, A. Gerasoulis, "DSC: Scheduling Parallel Tasks on an Unbounded Number of Processors," IEEE Transactions on Parallel and Distributed Systems, vol. 5, no. 9, pp. 951967, September, 1994.  
BibTex  x  
@article{ 10.1109/71.308533, author = {T. Yang and A. Gerasoulis}, title = {DSC: Scheduling Parallel Tasks on an Unbounded Number of Processors}, journal ={IEEE Transactions on Parallel and Distributed Systems}, volume = {5}, number = {9}, issn = {10459219}, year = {1994}, pages = {951967}, doi = {http://doi.ieeecomputersociety.org/10.1109/71.308533}, publisher = {IEEE Computer Society}, address = {Los Alamitos, CA, USA}, }  
RefWorks Procite/RefMan/Endnote  x  
TY  JOUR JO  IEEE Transactions on Parallel and Distributed Systems TI  DSC: Scheduling Parallel Tasks on an Unbounded Number of Processors IS  9 SN  10459219 SP951 EP967 EPD  951967 A1  T. Yang, A1  A. Gerasoulis, PY  1994 KW  Index Termsscheduling; directed graphs; computational complexity; parallel algorithms; trees(mathematics); parallel programming; DSC; parallel task scheduling; lowcomplexityheuristic; dominant sequence clustering algorithm; completely connected processor;unbounded number; performance; nonzero communication overhead; arbitrary directedacyclic task graphs; DAGs; NPcomplete; optimal schedules; special classes; fork; join;coarsegrain trees; finegrain trees; general scheduling algorithms; ETF; MD VL  5 JA  IEEE Transactions on Parallel and Distributed Systems ER   
We present a lowcomplexity heuristic, named the dominant sequence clusteringalgorithm (DSC), for scheduling parallel tasks on an unbounded number of completelyconnected processors. The performance of DSC is on average, comparable to, or evenbetter than, other highercomplexity algorithms. We assume no task duplication andnonzero communication overhead between processors. Finding the optimum solution forarbitrary directed acyclic task graphs (DAG's) is NPcomplete. DSC finds optimalschedules for special classes of DAG's, such as fork, join, coarsegrain trees, and somefinegrain trees. It guarantees a performance within a factor of 2 of the optimum forgeneral coarsegrain DAG's. We compare DSC with three highercomplexity generalscheduling algorithms: the ETF by J.J. Hwang, Y.C. Chow, F.D. Anger, and C.Y. Lee(1989); V. Sarkar's (1989) clustering algorithm; and the MD by M.Y. Wu and D. Gajski(1990). We also give a sample of important practical applications where DSC has beenfound useful.
[1] M. A. AlMouhamed, "Lower bound on the number of processors and time for scheduling precedence graphs with communication costs,"IEEE Trans. Software Eng., vol. 16, pp. 13901401, 1990.
[2] F. D. Anger, J. Hwang, and Y. Chow, "Scheduling with sufficient loosely coupled processors,"J. Parallel and Distributed Comput., vol. 9, pp. 8792, 1990.
[3] P. Chretienne, "Task scheduling over distributed memory machines," inProc. Int. Workshop Parallel Distrib. Algorithms, 1989.
[4] P. Chretienne, "A polynomial algorithm to optimially schedule tasks over an ideal distributed system under treelike presedence constraints,"European J. Oper. Res., vol. 2, pp. 225230, 1989.
[5] P. Chretienne, "Complexity of tree scheduling with interprocessor communication delays," Tech. Rep. M.A.S.I. 90.5, Universite Pierre et Marie Curie, 1990.
[6] J. Y. Colin and P. Chretienne, "C. P. M. scheduling with small communication delays and task duplication,"Oper. Res., vol. 39, no. 4, pp. 680684, 1991.
[7] M. Cosnard, M. Marrakchi, Y. Robert, and D. Trystram, "Parallel Gaussian elimination on an MIMD computer,"Parallel Computing, vol. 6, pp. 275296, 1988.
[8] A. Gerasoulis and T. Yang, "On the granularity and clustering of directed acyclic task graphs,"IEEE Trans. Parallel Distrib. Syst., vol. 4, pp. 686701, June 1993.
[9] A. Gerasoulis and T. Yang, "A comparison of clustering heuristics for scheduling DAG's on multiprocessors,"J. Parallel Distrib. Computing, vol. 16, pp. 276291, Dec. 1992.
[10] M. Girkar and C. Polychronopoulos, "Partitioning programs for parallel execution," inProc. 1988 ACM Int. Conf. Supercomputing, 1988.
[11] J.J. Hwang, Y.C. Chow, F. D. Anger, and C. Y. Lee, "Scheduling precedence graphs in systems with interprocessor communication times,"SIAM Computing, pp. 244257, Apr. 1988.
[12] S. J. Kim and J. C. Browne, "A general approach to mapping of parallel computation upon multiprocessor architectures," inInt. Conf. Parallel Processing, vol. 3, pp. 18, 1988.
[13] B. Kruatrachue and T. Lewis, "Grain size determination for parallel processing,"IEEE Software, Jan. 1988, pp. 2332.
[14] C. McCreary and H. Gill, "Automatic determination of grain size for efficient parallel processing,"Commun. ACM, vol. 32, pp. 10731078, Sept., 1989.
[15] C. Papadimitriou and M. Yannakakis, "Toward an architectureindependent analysis of parallel algorithms,"SIAM J. Comput., vol. 19, no. 2, pp. 322328, Apr. 1990.
[16] R. Pozo, "Performance modeling of sparse matrix methods for distributed memory architectures," inLecture Notes in Computer Science 634, Parallel Processing: CONPAR 92VAPPV. New York: SpringerVerlag, 1992, pp. 677688.
[17] V. Sarkar,Partitioning and Scheduling Parallel Programs for Multiprocessing, MIT Press, 1989.
[18] R. Wolski and J. Feo, "Program parititoning for NUMA multiprocessor computer systems,"J. Parallel Distrib. Computing(special issue on performance of supercomputers), vol. 19, pp. 203218, 1993.
[19] M. Y. Wu and D. Gajski, "Hypertool: A programming aid for messagepassing systems,"IEEE Trans. Parallel Distrib. Syst., vol. 1, pp. 330343, 1990.
[20] M. Y. Wu, personal commun., Feb. 1993.
[21] J. Yang, L. Bic, and A. Nicolau, "A mapping strategy for MIMD computers,"Proc. 1991 Int. Conf. Parallel Processing, vol. I, pp. 102109.
[22] T. Yang and A. Gerasoulis, "A fast static scheduling algorithm for DAG's on an unbounded number of processors," inProc. IEEE Supercomputing '91, IEEE, Albuquerque, NM, Nov. 1991, pp. 633642.
[23] T. Yang and A. Gersoulis, "Pyrros: Static Task Scheduling and Code Generation for MessagePassing Multiprocessors,"Proc. 6th ACM Int'l Conf. Supercomputing, ACM Press, New York, 1992, pp. 428443.