[openib-general] Announcing the release of MVAPICH2 0.9.2 (MPI-2 for InfiniBand and other RDMA Interconnects)

Dhabaleswar Panda panda at cse.ohio-state.edu
Sun Jan 15 18:29:14 PST 2006


The MVAPICH team is pleased to announce the release of MVAPICH2 0.9.2
with OpenIB/Gen2, VAPI and uDAPL transport interfaces. It also has
support for the standard TCP/IP (provided by MPICH2 stack). It is
optimized for the following platforms, OS, compilers and InfiniBand
adapters:

  - Platforms: EM64T, Opteron, IA-32 and Mac G5 
  - Operating Systems: Linux, Solaris and Mac OSX 
  - Compilers: gcc, intel, pathscale and pgi 
  - InfiniBand Adapters: Mellanox adapters with PCI-X 
    and PCI-Express (SDR and DDR with mem-full and mem-free cards) 

Starting with this release, MVAPICH2 enables InfiniBand support for
OpenIB/Gen2. All features available for the VAPI and uDAPL interfaces
are now available for the OpenIB/Gen2 interface.

MVAPICH2 0.9.2 is being distributed as a single integrated package
(with MPICH2 1.0.2p1 and MVICH).  It is available under BSD license.

This new release has the following features:

      - single code base with multiple underlying transport
        interfaces: VAPI, OpenIB/Gen2, uDAPL and TCP/IP
      - high-performance and optimized support for many MPI-2
        functionalities (one-sided, collectives, datatype)
      - support for other MPI-2 functionalities (as provided by 
        MPICH2 1.0.2p1). 
      - high-performance and optimized support for all MPI-1 
        functionalities (including two-sided)
      - high performance and optimized support for all one-sided 
        operations (Get, Put, and Accumulate)
      - support for both active and passive synchronization
      - optimized two-sided operations with RDMA support
      - efficient memory registration/de-registration schemes 
        for RDMA operations
      - optimized intra-node shared memory support (bus-based and
        NUMA)
      - shared library support for existing binary MPI programs to run
      - ROMIO support for MPI-IO
      - uDAPL support for portability across networks and OS 
        (tested for InfiniBand on Linux and Solaris; and Myrinet) 
      - scalable job start-up
      - optimized and tuned for the above platforms and different
        network interfaces (PCI-X and PCI-Express with SDR and DDR)
      - support for multiple compilers (gcc, icc, pathscale and pgi) 
      - single code base for all of the above platforms and OS
      - integrated and easy-to-use build script for installing the
        code on various platforms, OS, compilers, devices, and
        InfiniBand adapters
      - incorporates a set of runtime and compiler time tunable
        parameters for convenient tuning on large-scale clusters 

Other features of this release include:

- Excellent performance: Sample performance numbers include:
 
  - OpenIB/Gen2 on EM64T with PCI-Ex and IBA-DDR:
      Two-sided operations: 
        - 3.08 microsec one-way latency (4 bytes)
        - 1476 MB/sec unidirectional bandwidth 
        - 2661 MB/sec bidirectional bandwidth  

      One-sided operations:
        - 4.84 microsec Put latency 
        - 1483 MB/sec unidirectional Put bandwidth 
        - 2661 MB/sec bidirectional Put bandwidth 

  - OpenIB/Gen2 on EM64T with PCI-Ex and IBA-SDR:
      Two-sided operations:
        - 3.35 microsec one-way latency (4 bytes) 
        - 964 MB/sec unidirectional bandwidth
        - 1846 MB/sec bidirectional bandwidth 

      One-sided operations:
        - 5.43 microsec Put latency
        - 964 MB/sec unidirectional Put bandwidth
        - 1846 MB/sec bidirectional Put bandwidth

  - OpenIB/Gen2 on Opteron with PCI-Ex and IBA-SDR: 
      Two-sided operations:
        - 3.27 microsec one-way latency (4 bytes) 
        - 968 MB/sec unidirectional bandwidth
        - 1896 MB/sec bidirectional bandwidth 

      One-sided operations:
        - 5.95 microsec Put latency
        - 968 MB/sec unidirectional Put bandwidth
        - 1896 MB/sec bidirectional Put bandwidth

  - Solaris uDAPL/IBTL on Opteron with PCI-X and IBA-SDR:
      Two-sided operations:
        - 5.58 microsec one-way latency (4 bytes)
        - 655 MB/sec unidirectional bandwidth
        - 799 MB/sec bidirectional bandwidth

  - OpenIB/Gen2 uDAPL on Opteron with PCI-Ex and IBA-SDR:
      Two-sided operations:
        - 3.63 microsec one-way latency (4 bytes)
        - 962 MB/sec unidirectional bandwidth
        - 1869 MB/sec bidirectional bandwidth

  Performance numbers for all other platforms, system configurations
  and operations can be viewed by visiting `Performance Results'
  section of the project's web page.

- Similar performance with MVAPICH: With the new ADI-3-level design,
  MVAPICH2 0.9.2 delivers similar performance for two-sided operations
  compared to MVAPICH 0.9.6. Organizations and users interested in
  getting the best performance for both two-sided and one-sided 
  operations may migrate from MVAPICH code base to MVAPICH2 code base. 

- A set of benchmarks to evaluate both two-sided and one-sided
  operations (Put, Get, and Accumulate)

- An enhanced and detailed `User Guide' to assist users: 

       - to install this package on different platforms
            with interfaces (VAPI, uDAPL, OpenIB/Gen2 and TCP/IP) and
            different options

       - to vary different parameters of the MPI installation to 
            extract maximum performance and achieve scalability,
            especially on large-scale systems.

You are welcome to download the MVAPICH2 0.9.2 package and access
relevant information from the following URL:

http://nowlab.cse.ohio-state.edu/projects/mpi-iba/

A stripped down version of this release is available at the OpenIB SVN. 

A successive version with additional features and integrated with
MPICH2 1.0.3 will be available soon.

All feedbacks, including bug reports and hints for performance tuning,
are welcome. Please send an e-mail to mvapich-help at cse.ohio-state.edu.

Thanks, 

MVAPICH Team at OSU/NBCL 

----------

PS: If you would like to be removed from this mailing list, please end
an e-mail to mvapich_request at cse.ohio-state.edu.


======================================================================
MVAPICH/MVAPICH2 project is currently supported with funding from
U.S. National Science Foundation, U.S. DOE Office of Science,
Mellanox, Intel, Cisco Systems, Sun Microsystems and Linux Networx;
and with equipment support from AMD, Apple, Appro, IBM, Intel,
Mellanox, Microway, PathScale, SilverStorm and Sun Microsystems. Other
technology partner includes Etnus.
======================================================================




More information about the general mailing list