Solving Sparse Least Squares Problems on Massively Distributed Memory Computers

نویسنده

  • Laurence T. Yang
چکیده

In this paper we study the parallel aspects of PCGLS, a basic iterative method whose main idea is to organize the computation of conjugate gradient method with preconditioner applied to normal equations, and Incomplete Modified Gram-Schmidt (IMGS) preconditioner for solving sparse least squares problems on massively parallel distributed memory computers. The performance of these methods on this kind of architecture is always limited because of the global communication required for the inner products. We will describe the parallelization of PCGLS and IMGS preconditioner by two ways of improvement. One is to assemble the results of a number of inner products collectively and the other is to create situations where communication can be overlapped with computation. A theoretical model of computation and communication phases is presented which allows us to decide the number of processors that minimizes the runtime. Several numerical experiments on Parsytec GC/PowerPlus are presented.

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Data Distribution Analysis of MCGLS Algorithm for Parallel Least Squares Problems

In this paper we mainly study diierent data distribution of MCGLS, a modiied algorithm of CGLS which is a basic iterative method to organize the computation of conjugate gradient method applied to normal equations, for solving sparse least squares problems on massively parallel distributed memory computers. The performance of CGLS on this kind of architecture is always limited because of the gl...

متن کامل

Scalability Analysis of CGLS Algorithm for Sparse Least Squares Problems on Massively Distributed Memory Computers

In this paper we study the parallelization of CGLS, a basic iterative method for large and sparse least squares problems whose main idea is to organize the computation of conjugate gradient method to normal equations. A performance model of computation and communication phases with isoeeciency concept are used to analyze the qualitative scalability behavior of this method implemented on massive...

متن کامل

Isoefficiency Analysis of CGLS Algorithms for Parallel Least Squares Problems

In this paper we study the parallelization of CGLS, a basic iterative method for large and sparse least squares problems whose main idea is to organize the computation of conjugate gradient method to normal equations. A performance model called isoeeciency concept is used to analyze the behavior of this method implemented on massively parallel distributed memory computers with two dimensional m...

متن کامل

Parallel Execution Time Analysis for Least Squares Problems on Distributed Memory Architectures

In this paper we study the parallelization of PCGLS, a basic iterative method which main idea is to organize the computation of conjugate gradient method with preconditioner applied to normal equations. Two important schemes are discussed. What is the best possible data distribution and which communication network topology is most suitable for solving least squares problems on massively paralle...

متن کامل

The Highly Parallel Incomplete Gram-Schmidt Preconditioner

In this paper we study the parallel aspects of IMGS, Incomplete Modiied Gram-Schmidt preconditioner which can be used for ef-ciently solving sparse and large linear systems and least squares problems on massively parallel distributed memory computers. The performance of this preconditioning technique on this kind of architecture is always limited because of the global communication required for...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 1997