Hi,I have just installed intel mpi (v4.0.0.027) on a nahelem-IB based cluster system which uses the SLURM resource manager. All of the compilers and mpi stacks are installed using modules, including the intel mpi. After I load the intel-mpi module, build the application, and try to run it using a SLURM batch file, the program crashes, as the intel-mpi runtime environment does not obtain all of the SLURM environments. I get the message..<---mpiexec_rm1867: cannot connect to local mpd (/tmp/mpd2.console_apurkay); possible causes:1. no mpd is running on this host2. an mpd is running but was started without a "console" (-n option):--->It does not get the information on the compute nodes and tries instead to run on the login-console which is not the place to run, and hence fails.I assume then that the SLURM environments relating to the mpd.hosts file was not captured by the intel-mpi? If so, what runtime parameters or environments do need to pass/define in the slurm batch script?BTW, the default setup with OMPI/OpenFabrics and SLURM works fine.Thanks for any help.-- Avi
For more complete information about compiler optimizations, see our Optimization Notice.