DocumentCode :
1405574
Title :
A framework for exploiting task and data parallelism on distributed memory multicomputers
Author :
Ramaswamy, Shankar ; Sapatnekar, Sachin ; Banerjee, Prithviraj
Author_Institution :
Transarc Corp., Pittsburgh, PA, USA
Volume :
8
Issue :
11
fYear :
1997
fDate :
11/1/1997 12:00:00 AM
Firstpage :
1098
Lastpage :
1116
Abstract :
Distributed Memory Multicomputers (DMMs), such as the IBM SP-2, the Intel Paragon, and the Thinking Machines CM-5, offer significant advantages over shared memory multiprocessors in terms of cost and scalability. Unfortunately, the utilization of all the available computational power in these machines involves a tremendous programming effort on the part of users, which creates a need for sophisticated compiler and run-time support for distributed memory machines. In this paper, we explore a new compiler optimization for regular scientific applications-the simultaneous exploitation of task and data parallelism. Our optimization is implemented as part of the PARADIGM HPF compiler framework we have developed. The intuitive idea behind the optimization is the use of task parallelism to control the degree of data parallelism of individual tasks. The reason this provides increased performance is that data parallelism provides diminishing returns as the number of processors used is increased. By controlling the number of processors used for each data parallel task in an application and by concurrently executing these tasks, we make program execution more efficient and, therefore, faster
Keywords :
distributed memory systems; optimising compilers; parallel programming; CM-5; DMMs; IBM SP-2; Intel Paragon; PARADIGM HPF compiler framework; Thinking Machines; allocation; compiler optimization; convex programming; data parallel; data parallelism; distributed memory; distributed memory multicomputers; program execution; run-time support; scheduling; task parallel; task parallelism; Concurrent computing; Costs; Data structures; Distributed computing; Optimizing compilers; Parallel processing; Program processors; Random access memory; Runtime; Scalability;
fLanguage :
English
Journal_Title :
Parallel and Distributed Systems, IEEE Transactions on
Publisher :
ieee
ISSN :
1045-9219
Type :
jour
DOI :
10.1109/71.642945
Filename :
642945
Link To Document :
بازگشت