Compiler Algorithms for Optimizing Locality and Parallelism on Shared and Distributed-Memory Machines
Document Type
Article
Publication Date
8-1-2000
Abstract
Distributed-memory message-passing machines deliver scalable performance but are difficult to program. Shared-memory machines, on the other hand, are easier to program but obtaining scalable performance with large number of processors is difficult. Recently, scalable machines based on logically shared physically distributed memory have been designed and implemented. While some of the performance issues like parallelism and locality are common to different parallel architectures, issues such as data distribution are unique to specific architectures. One of the most important challenges compiler writers face is the design of compilation techniques that can work well on a variety of architectures. In this paper, we propose an algorithm that can be employed by optimizing compilers for different types of parallel architectures. Our optimization algorithm does the following: (1) transforms loop nests such that, where possible, the iterations of the outermost loops can be run in parallel across processors; (2) optimizes memory locality by carefully distributing each array across processors; (3) optimizes interprocessor communication using message vectorization whenever possible; and (4) optimizes cache locality by assigning appropriate storage layout for each array and by transforming the iteration space. Depending on the machine architecture, some or all of these steps can be applied in a unified framework. We report empirical results on an SGI Origin 2000 distributed-shared-memory multiprocessor and an IBM SP-2 distributed-memory message-passing machine to validate the effectiveness of our approach. © 2000 Academic Press.
Publication Source (Journal or Book title)
Journal of Parallel and Distributed Computing
First Page
924
Last Page
965
Recommended Citation
Kandemir, M., Ramanujam, J., & Choudhary, A. (2000). Compiler Algorithms for Optimizing Locality and Parallelism on Shared and Distributed-Memory Machines. Journal of Parallel and Distributed Computing, 60 (8), 924-965. https://doi.org/10.1006/jpdc.2000.1639