Using Multiple DAPL* Providers with the Intel® MPI Library


If your MPI program sends messages of drastically different sizes (for example, some 16 byte messages, and some 4 megabyte messages), you want optimum performance at all message sizes.  This cannot easily be obtained with a single DAPL* provider.  This is due to latency being a major factor for smaller messages, but bandwidth being more important for larger messages, and providers often making a tradeoff to improve one or the other.  The Intel® MPI Library, as of Version 4.1 Update 1, now supports up to three providers for a single rank of your job.


The ideal scenario is to allow the automatic provider detection to choose the providers for your job.  The automatic detection has additional features for finding the appropriate adapter to use in systems with multiple adapters, and can easily set each rank to use different adapters.  If you want to control the provider selection directly, you can do so with the environment variable I_MPI_DAPL_PROVIDER_LIST.  This takes a comma separated list of up to three providers with the syntax I_MPI_DAPL_PROVIDER_LIST=<Provider1,Provider2,Provider3>, where:

  • Provider1 is used for small messages.  This provider should be a low latency provider.
  • Provider2 is used for large messages inside of a single node.  This is most applicable in jobs involving Intel® Xeon Phi™ Coprocessors, or if you are using DAPL* for intranode communications.
  • Provider3 is used for large messages between multiple nodes.

 The following environment variables can also take a comma separated list of options.  The first option will be applied to the first provider, the second to the second, and the third to the third.  If only one value is set, it is applied to all providers, and if an invalid value is given, all providers will fall back to default values.



For more complete information about compiler optimizations, see our Optimization Notice.