Chapter 1. Introduction

Message Passing Toolkit (MPT) is a software package that supports interprocess data exchange for applications that use concurrent, cooperating processes on a single host or on multiple hosts. Data exchange is done through message passing, which is the use of library calls to request data delivery from one process to another or between groups of processes.

The MPT package contains the following components and the appropriate accompanying documentation:

SGI MPT is highly optimized for all SGI hardware platforms. The SGI ProPack 7 for Linux Start Here lists all current SGI software and hardware manuals and can be found on the SGI Technical Publications Library at: http://docs.sgi.com.

This chapter provides an overview of the MPI software that is included in the toolkit. It includes descriptions of the MPI standard compliance, the basic components of MPI, and the basic features of MPI. Subsequent chapters address the following topics:

MPI Overview

MPI was created by the Message Passing Interface Forum (MPIF). MPIF is not sanctioned or supported by any official standards organization. Its goal was to develop a widely used standard for writing message passing programs.

SGI supports implementations of MPI that are released as part of the Message Passing Toolkit. The MPI Standard is documented online at the following address:http://www.mcs.anl.gov/mpi

MPI 2.0 Standard Compliance

The SGI MPI implementation is compliant with the MPI 2.0 standard.

MPI Components

The MPI library is provided as a dynamic shared object (DSO) (a file with a name that ends in .so). The basic components that are necessary for using MPI are the libmpi.so library, the include files, and the mpirun command.

Profiling support is included in the libmpi.so library. Profiling support replaces all MPI_Xxx prototypes and function names with PMPI_Xxx entry points.

SGI MPI Features

The SGI MPI implementation offers a number of significant features that make it the preferred implementation to use on SGI hardware:

  • Data transfer optimizations for NUMAlink where available, including single-copy data transfer

  • Multi-rail InfiniBand support to take full advantage of the multiple InfiniBand fabrics available on SGI Altix ICE systems

  • Use of hardware fetch operations (fetchops), where available, for fast synchronization and lower latency for short messages

  • Optimized MPI-2 one-sided commands

  • Interoperability with the SHMEM (LIBSMA) programming model

  • High performance communication support for partitioned systems via XPMEM

  • Internet Wide Area RDMA Protocol (iWARP) support to take advantage of the Remote DMA capabilities of 10-Gigabit Ethernet