MPI-Support

Intel MPI Library and Composer XE Compatibility

The following table lists all supported versions of the Intel® MPI Library and the Intel® Composer XE.  Use this as a reference on the cross-compatibility between the library and associated compiler.

  • Linux*
  • Microsoft Windows* (XP, Vista, 7)
  • Microsoft Windows* 8
  • C/C++
  • Fortran
  • Intel® C++ Composer XE
  • Intel® Composer XE
  • Intel® Fortran Compiler
  • Intel® Fortran Composer XE
  • Intel® Visual Fortran Composer XE
  • Intel® MPI Library
  • Intel® Cluster Studio XE
  • MPI
  • compiler
  • compatibility
  • version
  • Message Passing Interface
  • Cluster-Computing
  • Intel® MPI Library 5.0 Beta Update 1 Readme

    The Intel® MPI Library is a high-performance interconnect-independent multi-fabric library implementation of the industry-standard Message Passing Interface, v3.0 (MPI-3.0) specification. This Beta package is for MPI users who develop on and build for Intel® 64 architectures on Linux* and Windows*, as well as customers running on the Intel® Xeon Phi™ coprocessor on Linux*. You must have a valid license to download, install and use this product.

  • Linux*
  • Microsoft Windows* (XP, Vista, 7)
  • Microsoft Windows* 8
  • Server
  • C/C++
  • Fortran
  • Intel® MPI Library
  • Message Passing Interface
  • Cluster-Computing
  • Using Intel MPI over Rdma over Converged Ethernet

    Hello All,

    I have been trying to run IMB over 40GigE RoCE using IntelMPI and OpenMPI. I could run it using OpenMPI using the following FAQ 

    https://www.open-mpi.org/faq/?category=openfabrics#ompi-over-roce.

    But I have been having issues running IMB using Intel MPI. I did not see many resources online. I have been trying to run it as follows.

    Intel® MPI Library 4.1 Update 3 Build 049 Readme

    The Intel® MPI Library for Linux* and Windows* is a high-performance interconnect-independent multi-fabric library implementation of the industry-standard Message Passing Interface, v2.2 (MPI-2.2) specification. This package is for MPI users who develop on and build for IA-32 and Intel® 64 architectures on Linux* and Windows*, as well as customers running on the Intel® Xeon Phi™ coprocessor on Linux*. You must have a valid license to download, install and use this product.

  • Linux*
  • Server
  • C/C++
  • Fortran
  • Intel® MPI Library
  • Message Passing Interface
  • Cluster-Computing
  • Intel® MPI Library 4.1 Update 3 Build 048 Readme

    The Intel® MPI Library for Linux* and Windows* is a high-performance interconnect-independent multi-fabric library implementation of the industry-standard Message Passing Interface, v2.2 (MPI-2.2) specification. This package is for MPI users who develop on and build for IA-32 and Intel® 64 architectures on Linux* and Windows*, as well as customers running on the Intel® Xeon Phi™ coprocessor on Linux*. You must have a valid license to download, install and use this product.

  • Linux*
  • Microsoft Windows* (XP, Vista, 7)
  • Microsoft Windows* 8
  • Server
  • C/C++
  • Fortran
  • Intel® MPI Library
  • Message Passing Interface
  • Cluster-Computing
  • Dual-rail MPI binding

    Dear expert,

    I seek confirmation that I am doing stuff properly. Here my situation. The new cluster in my institution has two Mellanox Connect-IB cards on each node. Each node is a dual socket six-core Ivy Bridge. The node architecture is such that each socket is connected with a straight PCIe lane to each IB card. What I want to do is basically assign a subset of the MPI processes (e.g. the first 6) to first IB card and the other MPI processes to the second IB card. No rail sharing, for both small and large messages a MPI should use one single (assigned) IB card.

    Intel® MPI Library 4.1 Update 3 Readme

    The Intel® MPI Library for Linux* and Windows* is a high-performance interconnect-independent multi-fabric library implementation of the industry-standard Message Passing Interface, v2.2 (MPI-2.2) specification. This package is for MPI users who develop on and build for IA-32 and Intel® 64 architectures on Linux* and Windows*, as well as customers running on the Intel® Xeon Phi™ coprocessor on Linux*. You must have a valid license to download, install and use this product.

  • Entwickler
  • Linux*
  • Microsoft Windows* (XP, Vista, 7)
  • Server
  • C/C++
  • Fortran
  • Intel® MPI Library
  • Message Passing Interface
  • Cluster-Computing
  • Intel® MPI Library 4.1 Update 2 Readme

    The Intel® MPI Library for Linux* and Windows* is a high-performance interconnect-independent multi-fabric library implementation of the industry-standard Message Passing Interface, v2.2 (MPI-2.2) specification. This package is for MPI users who develop on and build for IA-32 and Intel® 64 architectures on Linux* and Windows*, as well as customers running on the Intel® Xeon Phi™ coprocessor on Linux*. You must have a valid license to download, install and use this product.

  • Linux*
  • Microsoft Windows* (XP, Vista, 7)
  • Microsoft Windows* 8
  • Server
  • C/C++
  • Fortran
  • Intel® MPI Library
  • Message Passing Interface
  • Cluster-Computing
  • MPI-Support abonnieren