intel MPI codes fails with more than 1 node

intel MPI codes fails with more than 1 node

For a simple mpi program compiled with Intel compiler studio 2016 and 2017, with intel compiler and mpi, the jobs fail with the following debug errors. The code will run extremely slowly, and get stuck for about 30 seconds at one stage if run on 2 or more nodes. It runs smoothly without any problems on a single node. The same code compiled with gcc and openmpi runs smoothly without any problem on any number of nodes.

Do you know what might be the problem? Thanks.

AttachmentSize
Downloadapplication/octet-stream error.log35.2 KB
3 posts / 0 new
Last post
For more complete information about compiler optimizations, see our Optimization Notice.

This question might get more expert attention on the companion cluster/hpc forum https://software.intel.com/en-us/forums/intel-clusters-and-hpc-technology

Your error.log appears to be the results from a successful "Hello World" test program.

>>The code will run extremely slowly, and get stuck for about 30 seconds at one stage if run on 2 or more nodes

Sounds like a coding error on your part. Handshake issues resulting in deadlock.

Jim Dempsey

Leave a Comment

Please sign in to add a comment. Not a member? Join today