Global Hydra Options
This section describes the global options of the Intel® MPI Library's Hydra process manager. Global options are applied to all arguments sets in the launch command. Argument sets are separated by a colon '
:
'.
-tune
<filename>
Use this option to specify the file name that contains the tuning data in a binary format.
-usize
<usize>
Use this option to set
MPI_UNIVERSE_SIZE
, which is available as an attribute of the
MPI_COMM_WORLD
.
<size> | Define the universe size
|
SYSTEM | Set the size equal to the number of cores passed to
mpiexec through the hostfile or the resource manager.
|
INFINITE | Do not limit the size. This is the default value.
|
<value> | Set the size to a numeric value ≥ 0.
|
-hostfile
<hostfile>
or -f
<hostfile>
Use this option to specify host names on which to run the application. If a host name is repeated, this name is used only once.
See also the
I_MPI_HYDRA_HOST_FILE
environment variable for more details.
Use the following options to change the process placement on the cluster nodes:
-machinefile
<machine file>
or -machine
<machine file>
Use this option to control process placement through a machine file. To define the total number of processes to start, use the
-n
option. For example:
$ cat ./machinefile node0:2 node1:2 node0:1
-hosts-group
Use this option to set node ranges using brackets, commas, and dashes (like in Slurm* Workload Manager).
For more details, see the
I_MPI_HYDRA_HOST_FILE
environment variable in
Hydra Environment Variables.
-silent-abort
Use this option to disable abort warning messages.
For more details, see the
I_MPI_SILENT_ABORT
environment variable in
Hydra Environment Variables.
-nameserver
Use this option to specify the nameserver in the
hostname:port
format.
For more details, see the
I_MPI_HYDRA_NAMESERVER
environment variable in
Hydra Environment Variables.
-genv
<ENVVAR> <value>
Use this option to set the
environment variable to the specified
for all MPI processes.
<ENVVAR>
<value>
-genvall
Use this option to enable propagation of all environment variables to all MPI processes.
-genvnone
Use this option to suppress propagation of any environment variables to any MPI processes.
The option does not work for localhost.
-genvexcl
<list of env var names>
Use this option to suppress propagation of the listed environment variables to any MPI processes.
-genvlist
<list>
Use this option to pass a list of environment variables with their current values.
is a comma separated list of environment variables to be sent to all MPI processes.
<list>
-pmi-connect
<mode>
Use this option to choose the caching mode of process management interface (PMI) message. Possible values for
<mode>
are:
<mode> | The caching mode to be used
|
nocache | Do not cache PMI messages.
|
cache | Cache PMI messages on the local
pmi_proxy management processes to minimize the number of PMI requests. Cached information is automatically propagated to child management processes.
|
lazy-cache | cache mode with on-request propagation of the PMI information.
|
alltoall | Information is automatically exchanged between all
pmi_proxy before any get request can be done. This is the default mode.
|
See the
environment variable for more details.
-perhost
<# of processes >
, -ppn
<# of processes >,
or -grr
<# of processes>
Use this option to place the specified number of consecutive MPI processes on every host in the group using round robin scheduling. See the
environment variable for more details.
When running under a job scheduler, these options are ignored by default. To be able to control process placement with these options, disable the
variable.
-rr
Use this option to place consecutive MPI processes on different hosts using the round robin scheduling. This option is equivalent to "
-perhost 1
". See the
environment variable for more details.
-trace [
<profiling_library>
] or -t [<profiling_library>
]Use this option to profile your MPI application with Intel® Trace Collector using the indicated
. If you do not specify
, the default profiling library
<profiling_library>
<profiling_library>
libVT.so
is used.
Set the
I_MPI_JOB_TRACE_LIBS
environment variable to override the default profiling library.
-trace-imbalance
Use this option to profile your MPI application with Intel® Trace Collector using the
libVTim.so
library.
-aps
Use this option to collect statistics from your MPI application using Application Performance Snapshot. The collected data includes hardware performance metrics, memory consumption data, internal MPI imbalance and OpenMP* imbalance statistics. When you use this option, a new folder
aps_result_<date>-<time>
with statistics data is generated. You can analyze the collected data with the
aps
utility, for example:
$ mpirun -aps -n 2 ./myApp $ aps aps_result_20171231_235959
- To use this option, set up the Application Performance Snapshot environment beforehand. See the tool's https://software.intel.com/content/www/us/en/develop/documentation/application-snapshot-user-guide/top.htmlUser Guide.
- If you use the options-traceor-check_mpi, the-apsoption is ignored.
-mps
Use this option to collect only MPI and OpenMP* statistics from your MPI application using Application Performance Snapshot. Unlike the
-aps
option,
-mps
doesn't collect hardware metrics. The option is equivalent to:
$ mpirun -n 2 aps -c mpi,omp ./myapp
-trace-pt2pt
Use this option to collect the information about point-to-point operations using Intel® Trace Analyzer and Collector. The option requires that
you also use the
-trace
option.
-trace-collectives
Use this option to collect the information about collective operations using Intel® Trace Analyzer and Collector. The option requires that
you also use the
-trace
option.
Use the
-trace-pt2pt
and
-trace-collectives
to reduce the size of the resulting trace file or the number of message checker reports. These options work with both statically and dynamically linked applications.
-configfile
<filename>
Use this option to specify the file
that contains the command-line options with one executable per line. Blank lines and lines that start with '
<filename>
#
' are ignored. Other options specified in the command line are treated as global.
You can specify global options in configuration files loaded by default (
mpiexec.conf
in
<installdir>/etc
,
~/.mpiexec.conf
, and
mpiexec.conf
in the working directory). The remaining options can be specified in the command line.
-branch-count
<num>
Use this option to restrict the number of child management processes launched by the Hydra process manager, or by each
pmi_proxy
management process.
See the
I_MPI_HYDRA_BRANCH_COUNT
environment variable for more details.
-pmi-aggregate or -pmi-noaggregate
Use this option to switch on or off, respectively, the aggregation of the PMI requests. The default value is
-pmi-aggregate
, which means the aggregation is enabled by default.
See the
I_MPI_HYDRA_PMI_AGGREGATE
environment variable for more details.
-gdb
Use this option to run an executable under GDB* (GNU debugger). You can use the following command:
$ mpiexeс.hydra -gdb -n <# of processes> <executable>
-gdba
<pid>
Use this option to attach the GNU* debugger to the existing MPI job. You can use the following command:
$ mpiexec.hydra -gdba <pid>
-nolocal
Use this option to avoid running the
on the host where
<executable>
mpiexec
is launched. You can use this option on clusters that deploy a dedicated main node for starting the MPI jobs and a set of dedicated compute nodes for running the actual MPI processes.
.hydra
-hosts
<nodelist>
Use this option to specify a particular
on which the MPI processes should be run. For example, the following command runs the executable
on the hosts
and
:
<nodelist>
a.out
host1
host2
$ mpiexec.hydra -n 2 -ppn 1 -hosts host1,host2 ./a.out
If
<nodelist>
contains only one node, this option is interpreted as a local option. See
Local Options for details.
-iface
<interface>
Use this option to choose the appropriate network interface. For example, if the IP emulation of your InfiniBand* network is configured to
ib0
, you can use the following command.
$ mpiexec.hydra -n 2 -iface ib0 ./a.out
See the
I_MPI_HYDRA_IFACE
environment variable for more details.
-demux
<mode>
Use this option to set the polling mode for multiple I/O. The default value is
poll
.
Arguments
<spec> | Define the polling mode for multiple I/O
|
poll | Set
poll as the polling mode. This is the default value.
|
select | Set
select as the polling mode.
|
See the
I_MPI_HYDRA_DEMUX
environment variable for more details.
-enable-x or -disable-x
Use this option to control the Xlib* traffic forwarding. The default value is
-disable-x
, which means the Xlib traffic is not forwarded.
-l, -prepend-rank
Use this option to insert the MPI process rank at the beginning of all lines written to the standard output.
-ilp64
Use this option to preload the ILP64 interface.
-s
<spec>
Use this option to direct standard input to the specified MPI processes.
Arguments
<spec> | Define MPI process ranks
|
all
| Use all processes.
|
none | Do not direct standard output to any processes.
|
<l>,<m>,<n> | Specify an exact list and use processes
<l> <m> <n> |
<k>,<l>-<m>,<n> | Specify a range and use processes
<k> <l> <m> <n> |
-noconf
Use this option to disable processing of the
mpiexec.hydra
configuration files.
-ordered-output
Use this option to avoid intermingling of data output from the MPI processes. This option affects both the standard output and the standard error streams.
When using this option, end the last output line of each process with the end-of-line '
\n
' character. Otherwise the application may stop responding.
-path
<directory>
Use this option to specify the path to the executable file.
-tmpdir
<dir>
Use this option to set a directory for temporary files. See the
I_MPI_TMPDIR
environment variable for more details.
-version or -V
Use this option to display the version of the Intel® MPI Library.
-info
Use this option to display build information of the Intel® MPI Library. When this option is used, the other command line arguments are ignored.
-localhost
Use this option to explicitly specify the local host name for the launching node.
-rmk
<RMK>
Use this option to select a resource management kernel to be used. Intel® MPI Library only supports
pbs
.
See the
I_MPI_HYDRA_RMK
environment variable for more details.
-outfile-pattern
<file>
Use this option to redirect
stdout
to the specified file.
-errfile-pattern
<file>
Use this option to redirect
stderr
to the specified file.
-gpath
<path>
Use this option to specify the path to the executable file.
-gwdir
<dir>
Use this option to specify the working directory in which the executable file runs.
-gumask
<umask>
Use this option to perform the "
umask
" command for the remote executable file.
<umask>
-gdb-ia
Use this option to run processes under Intel® architecture specific GNU* debugger.
-prepend-pattern
Use this option to specify the pattern that is prepended to the process output.
-verbose or -v
Use this option to print debug information from
mpiexec
, such as:
.hydra
- Service processes arguments
- Environment variables and arguments passed to start an application
- PMI requests/responses during a job life cycle
See the
I_MPI_HYDRA_DEBUG
environment variable for more details.
-print-rank-map
Use this option to print out the MPI rank mapping.
-print-all-exitcodes
Use this option to print the exit codes of all processes.
-bootstrap
<bootstrap server>
Use this option to select a built-in bootstrap server to use. A bootstrap server is the basic remote node access mechanism that is provided by the system. Hydra supports multiple runtime bootstrap servers such as
ssh
,
rsh
,
pdsh
,
fork
,
persist
,
slurm
,
ll
,
lsf
, or
sge
to launch MPI processes. The default bootstrap server is
ssh
. By selecting
slurm
,
ll
,
lsf
, or
sge
, you use the corresponding
srun
,
llspawn.stdio
,
blaunch
, or
qrsh
internal job scheduler utility to launch service processes under the respective selected job scheduler (Slurm*, LoadLeveler*, LSF*, and SGE*).
Arguments
<arg> | String parameter
|
ssh | Use secure shell. This is the default value.
|
rsh | Use remote shell.
|
pdsh | Use parallel distributed shell.
|
pbsdsh | Use Torque* and PBS*
pbsdsh command.
|
fork | Use fork call. |
persist | Use Hydra persist server. See below for details.
|
slurm | Use Slurm*
srun command.
|
ll | Use LoadLeveler*
llspawn.stdio command.
|
lsf | Use LSF
blaunch command.
|
sge | Use Univa* Grid Engine*
qrsh command.
|
See
I_MPI_HYDRA_BOOTSTRAP
for details.
-bootstrap-exec
<bootstrap server>
Use this option to set the executable to be used as a bootstrap server. The default bootstrap server is
ssh
. For example:
$ mpiexec.hydra -bootstrap-exec <bootstrap_server_executable> -f hosts -env <VAR1> <VAL1> -n 2 ./a.out
See
I_MPI_HYDRA_BOOTSTRAP
for more details.
-bootstrap-exec-args <args>
Use this option to provide the additional parameters to the bootstrap server executable file.
$ mpiexec.hydra -bootstrap-exec-args <arguments> -n 2 ./a.out
For tight integration with the Slurm* scheduler (including support for suspend/resume), use the method outlined on the Slurm page here: http://www.schedmd.com/slurmdocs/mpi_guide.html#intel_mpi
See
I_MPI_HYDRA_BOOTSTRAP_EXEC_EXTRA_ARGS
for more details.
-v6
Use this option to force using the IPv6 protocol.