next up previous
Next: Parallel Efficiency Up: PerformancePortability, and Scalability Previous: Block Cyclic Data Layout

The BLACS as an Efficient, Portable, and Adequate Message-Passing Interface

The total volume of data communicated by most of the ScaLAPACK driver routines for dense matrices can be approximated by the quantity tex2html_wrap_inline674, where N is the order of the largest matrix operand. The number of messages, however, is proportional to N and can be approximated by the quantity tex2html_wrap_inline680, where nb is the logical blocking factor used in the computation. Similar to the situation described above, the ``standard'' constants tex2html_wrap_inline684 for the communication volume depend upon the performed computation and are of the same order as the floating-point-operation constants tex2html_wrap_inline628 shown in Table 4.

Developing an adequate message-passing interface specialized for linear algebra operations has been one of the first achievements of the ScaLAPACK project. The Basic Linear Algebra Communications Subprograms (BLACS) [19] were thus specifically designed to facilitate the expression of the relevant communication operations. The simplicity of the BLACS interface, as well as the rigor of their specification, allows for an easy port of the entire ScaLAPACK software. Currently, the BLACS have been efficiently ported on machine-specific message-passing libraries such as MPL (IBM) and NX (Intel), as well as more generic interfaces such as PVM and MPI. The BLACS overhead has been shown to be negligible [19].

The specificity and limited scope of the BLACS significantly contribute to its ease of use, portability, and efficiency. In addition, the BLACS interface provides the user and library designer with an appropriate level of notation. Indeed, the BLACS operate on typed two-dimensional arrays. The computational model consists of a one- or two-dimensional grid of processes, where each process stores matrices and vectors. The BLACS include synchronous send/receive routines to send a matrix or submatrix from one process to another, to broadcast submatrices, or to perform global reductions (sums, maxima and minima). Also included are routines to establish, modify, or query the process grid. The BLACS provide an adequate interface level for linear algebra communication operations.

For ease of use and flexibility, the BLACS send operation is locally blocking; that is, the return from the send operation indicates that the resources may be reused. However, since this feature depends only on local information, it is unknown whether the receive operation has been called. Buffering is necessary on the sending or the receiving process. The BLACS receive operation is globally blocking. The Return from the receive operation indicates that the message has been (sent and) received. On systems natively supporting globally blocking sends, nonblocking sends coupled with buffering are used to simulate locally blocking sends. This extra buffering operation may cause a slight performance degradation on those systems.

The BLACS broadcast and combine operations feature the ability to select different virtual network topologies. This easy-to-use built-in facility allows for the expression of various message scheduling approaches, such as a communication pipeline. This unique and distinctive BLACS characteristic is necessary for achieving the highest performance levels on distributed-memory platforms.


next up previous
Next: Parallel Efficiency Up: PerformancePortability, and Scalability Previous: Block Cyclic Data Layout

Jack Dongarra
Sat Feb 1 08:18:10 EST 1997