Expanding the Message Passing Library Model with Nested Parallelism
نویسنده
چکیده
A synchronous extension to the library model for message passing (Inmos C, PVM, Parmacs, MPI, etc.) is presented. This extension, provides a comfortable expression of nested parallelism from inside the message passing model. Furthermore of being a valuable tool for the presentation and teaching of parallel algorithms, the computational results prove that an efficiency similar to or even better than the one obtained designing and implementing algorithms using the native language can be achieved.
منابع مشابه
Supporting Coarse and Fine Grain Parallelism in an Extension of ML
We have built an extension of Standard ML aimed at multicomputer platforms with distributed memories. The resulting language, paraML, differs from other extensions by including and differentiating both coarse-grained and fine-grained parallelism. The basis for coarse-grained parallelism in paraML is process creation where there is no sharing of data, with communication between processes via asy...
متن کاملGAP/MPI: Facilitating parallelism
The goal of this work is to overcome the learning barriers faced when first using parallelism. Currently, in order to parallelize a system such as GAP, one must embed a message passing library such as MPI, with many routines and many parameters. GAP/MPI provides a simple, task-oriented interface sitting above the MPI library. The system presents the end-user with a single SPMD (single program, ...
متن کاملA Library-Based Approach to Task Parallelism in a Data-Parallel Language
Pure data-parallel languages such as High Performance Fortran version 1 (HPF) do not allow efficient expression of mixed task/data-parallel computations or the coupling of separately compiled data-parallel modules. In this paper, we show how these common parallel program structures can be represented, with only minor extensions to the HPF model, by using a coordination library based on the Mess...
متن کاملCluster-Based Parallel Image Processing
Many image processing tasks exhibit a high degree of data locality and parallelism and map quite readily to specialized massively parallel computing hardware. However, as workstation clusters are becoming a viable and economical parallel computing resource, it is important to understand how to use these environments for parallel image processing as well. In this paper we discuss our implementat...
متن کاملParallel and Distributed Algorithms for High Speed Image Processing
Many image processing tasks exhibit a high degree of data locality and parallelism and map quite readily to specialized massively parallel computing hardware. However, as distributed memory machines are becoming a viable and economical parallel computing resource, it is important to understand how to use these environments for parallel image processing as well. In this paper we discuss our impl...
متن کامل