Storage Mapping Optimization for Parallel Programs
نویسندگان
چکیده
Data dependences are known to hamper e cient parallelization of programs. Memory expansion is a general method to remove dependences in assigning distinct memory locations to dependent writes. Parallelization via memory expansion requires both moderation in the expansion degree and e ciency at run-time. We present a general storage mapping optimization framework for imperative programs, applicable to most loop nest parallelization techniques.
منابع مشابه
Parallelization via Constrained Storage Mapping Optimization
A key problem for parallelizing compilers is to nd the good tradeo between memory expansion and parallelism. This paper is a new step towards solving this problem. A framework for parallel execution order and storage mapping computation is designed, allowing time and space optimization. Constrained expansion|a theoretical model for expansion strategies|is shown to be very useful in this context.
متن کاملOptimization Flow for Algorithm Mapping on Graphics Cards
Graphics card architectures provide an optimal platform for parallel execution of many number crunching loop programs, ranging from fields like image processing to linear algebra. However, it is hard to efficiently map such algorithms to the graphics hardware, even with detailed insight into the architecture. This paper presents an optimization flow for mapping algorithms to the graphics hardwa...
متن کاملCommunication-conscious Mapping of Regular Nested Loop Programs onto Massively Parallel Processor Arrays
Methods for an efficient mapping of algorithms to parallel architectures are of utmost importance because many stateof-the-art embedded digital systems deploy parallelism to increase their computational power. This paper deals with the mapping of loop programs onto processor arrays implemented in an FPGA or available as (reconfigurable) coarsegrained processor architectures. Most existing work ...
متن کاملData Partitioning for a Good Node Performance
As a consequence of recent advances in interconnection network technology for MIMD parallel computers, optimizing communications in parallel programs has become a factor of secondary importance. For example, mapping processes onto processors is currently an issue of minor importance for some up-to-date distributed memory parallel computers, because the interconnection network guarantees a fairl...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
دوره شماره
صفحات -
تاریخ انتشار 1999