Intel® Clusters and HPC Technology

mpd error

Hi!

I have a problem with Altair PBS PRO + Intel MPI. I can launch a task with mpiexec command on several nodes. But when I try to launch this task on several nodes under PBS I get error.

What I doing:
1) Starting mpd on nodes:
qwer@mgr:/mnt/share/piex> cat mpd.hosts
ib-mgr:10
ib-cn01:16
ib-cn02:16
ib-cn03:16
ib-cn04:16
ib-cn05:16
qwer@mgr:/mnt/share/piex> mpdboot -n 6 -f mpd.hosts -r ssh

memory distribution

Hello everyone.
I'm want to run two mpi (MPICH2) codes in my cluster.
I send the first work distributed with round robin and all is ok.
The problem appear when I send the second work. The memory used by the second one is from the cpu used by the first work and doesn't use the memory of the other, almost free, cpu.
There is a flag with which I can tell to MPICH2 that it should use the memory free cpus?
Thank you.

Error in Intel MPI 3.2.2 MPI I/O

I've been testing the new Intel MPI 3.2.2 release which has support for Panasas' PanFS. I've checked out an evaluation copy of the library, but am running into what I believe is a bug with the use of layout hints to a shared file. For the case of more than 1 thread accessing a file, (N-to-1), this fails as each thread is trying to perform an ioctl call to the file and returning:

"ADIOI_PANFS_OPEN: I/O Error doing ioctl on parent directory to create PanFS file using ioctl: File exists."

mpdboot fails to start nodes with different users.

I am trying to figure out why a few nodes in my cluster are acting differently.

We are running Rocks 5.2 with RHEL 5
We use torque/maui as our queing system.
They submit jobs that use
MPI version 3.2.1.009

When I start a job as a user with this
mpdboot --rsh=ssh -d -v -n 16 -f /scr/username/testinput.nodes.mpd

Error with Intel MPI (ld: cannot find -lmpi)

I installed Intel MPI Library and tried compiling one of the test programs that come along with it (test.f). While doing so, here's the error message that I see.

[susheel@localhost intel_mpi_test]$ /opt/intel/impi/3.2.2.006/bin/mpiifort test.f
ld: skipping incompatible /opt/intel/impi/3.2.2.006/lib/libmpi.so when searching for -lmpi
ld: skipping incompatible /opt/intel/impi/3.2.2.006/lib/libmpi.a when searching for -lmpi
ld: cannot find -lmpi

MKL ScaLAPACK + MVAPICH + 100 lines of code = CRASH

The following code has proved good to generate a crash using MKL 10.2 update 2 (sequential version and threaded), last revision of MVAPICH, in two different clusters. Can anybody tell me what the problem is here? It does not crash always, but it does crash when the right number of MPI processes and matrix sizes are selected.

A

/*

* crash.cpp - crashes with ICC 11.1, MKL 10.2, MVAPICH 1.0 on linux 64-bit

* both linked with the serial or threaded libraries

* doing mpirun -np 36 crash 5000 10

*/

#include

#include

Iscriversi a Intel® Clusters and HPC Technology