Intel® Clusters and HPC Technology

Bug with -compile_info and -link_info options


When I run mpiicc -compile_info I get:

icc -I/softs/intel//impi/ -L/softs/intel//impi/ -L/softs/intel//impi/ -Xlinker --enable-new-dtags -Xlinker -rpath -Xlinker /softs/intel//impi/ -Xlinker -rpath -Xlinker /softs/intel//impi/ -Xlinker -rpath -Xlinker /opt/intel/mpi-rt/5.0/intel64/lib/release -Xlinker -rpath -Xlinker /opt/intel/mpi-rt/5.0/intel64/lib -lmpifort -lmpi -lmpigi -ldl -lrt -lpthread

hybrid application on the Xeon Phi

I would like to run a hybrid application (CP2K) on the Xeon Phi. The application is MPI + OpenMP and I set up the environment in the following manner: $ export OMP_NUM_THREADS=15 $ export I_MPI_PIN_PROCESSOR_LIST=$(seq -s "," 1 $OMP_NUM_THREADS 240) $ echo $I_MPI_PIN_PROCESSOR_LIST 1,16,31,46,61,76,91,106,121,136,151,166,181,196,211,226 $ mpirun -n $(expr 240 / $OMP_NUM_THREADS) However, the application is running awfully slowly. When I run the "top" command only shows the 16 MPI processes and not any of the threads and says the Phi system is 6.2% user busy (16 / 240 * 100).

mpitune -V ERROR

I installed the Intel Parallel Studio Cluster 15. 


The following command "impi_5.0.1/intel64/bin/tune/mpitune -V" shows ERROR:


There is nothing like /p/pdsd/Intel_MPI/Software/Python/python-2.7.2-linux-intel64-rhel5.7/ in our environment. Is it a setup error?


tune/mpitune  -V

ERROR:root:code for hash md5 was not found.

Traceback (most recent call last):

  File "/p/pdsd/Intel_MPI/Software/Python/python-2.7.2-linux-intel64-rhel5.7/lib/python2.7/", line 139, in <module>

Tracing MPI call with Intel MPI


Is there a way to trace MPI calls (meaning: each time an MPI function is called, I'd like to see the function name, parameters and emitting process) using Intel MPI ?

There is a --trace option, but it seems connected with Trace Analyzer which only seems able to print information regarding cumulated time spent in a given function ?

I've been trying to read the documentation for some time now, and it does not seems to be supported, which seems kind of strange...



Can shared memory work between processes running w/wo mpiexec?

I have a GUI.exe (with GUI) and engine.exe (without GUI). I am using shared memory for inter-process communication between these two exe. Everything was working fine before I use mpiexec. After I added mpiexec for engine.exe, these two processes cannot talk to each other through shared memory. It seems that the shared memory is "shield" by mpiexe and cannot be shared by outside world, since mpiexec itself is also using shared memory for communication for parallel computations.

By the way, these two exe run on the same PC.

Is there any way to overcome this problem?

qdel not killing all processes started under Intel MPI

Hi, when we run using Intel MPI with Hydra process manager (in a script submitted with qsub-- this is with OGS/GE 2011.11p1 on ROCKS 6.1 on a small blade cluster), qdel does not fully kill the job except when the the jobscript runs on the frontend. I have to kill the processes started by mpirun manually if the jobscript runs on a compute node. This is not a problem with OpenMPI.

Any ideas or suggestions on how to proceed with troubleshooting this would be much appreciated.

MPI Rank Binding

Hello all,

Intel MPI 4.1.3 on RHEL6.4: trying to bind ranks in two simple fashions:(a) 2 ranks to the same processor socket and (b) 2 ranks to different processor sockets.

Looking at the Intel MPI Reference Manual (3.2. Process Pinning pp.98+), we should be able to use options in mpiexec.hydra when the hostfile points to the same host

-genv I_MPI_PIN 1  -genv I_MPI_PIN_PROCESSOR_LIST all:bunch
-genv I_MPI_PIN 1  -genv I_MPI_PIN_PROCESSOR_LIST all:scatter


Subscribe to Intel® Clusters and HPC Technology