Abstract
Scheduling computations with communications is the theoretical basis for achieving efficient parallelism on distributed memory systems. We generalize Graham's task-level in a manner to incorporate the effects of computation and communication. A new scheduling is proposed by combining task priority with efficient management of processor idle time. We also propose an optimization called Iterative Refinement Scheduling (IRS) that iteratively schedules the forward and backward computation graph. The task-level used in some scheduling iteration is obtained from the schedule generated in the previous iteration. Each iteration produces a new schedule and new task-levels. This approach enables searching and optimizing solutions as the result of using more refined task-level in each scheduling iteration. Evaluation and analysis of the results are carried out for different instances of communication granularities and problem parallelism. It is shown that solutions obtained out of few iterations statistically outperforms those generated by other recently proposed scheduling. IRS allows exploring a space of solutions whose size grows with the amount of parallelism and communication granularity. IRS enables optimizing the solution specially for critical instances such as fine-grain computations and large parallelism.
| Original language | English |
|---|---|
| Pages (from-to) | 851-871 |
| Number of pages | 21 |
| Journal | Journal of Systems Architecture |
| Volume | 46 |
| Issue number | 10 |
| DOIs | |
| State | Published - Jul 2000 |
Bibliographical note
Funding Information:The authors acknowledge computing support and conference attendence support from the King Fahd University of Petroleum and Minerals, Dhahran 31261, Saudi Arabia.
Keywords
- Distributed memories
- Heuristics
- Message-passing
- Performance
- Scheduling
ASJC Scopus subject areas
- Software
- Hardware and Architecture