Parallel Computing

mpitune -V ERROR

I installed the Intel Parallel Studio Cluster 15. 

 

The following command "impi_5.0.1/intel64/bin/tune/mpitune -V" shows ERROR:

 

There is nothing like /p/pdsd/Intel_MPI/Software/Python/python-2.7.2-linux-intel64-rhel5.7/ in our environment. Is it a setup error?

 

tune/mpitune  -V

ERROR:root:code for hash md5 was not found.

Traceback (most recent call last):

  File "/p/pdsd/Intel_MPI/Software/Python/python-2.7.2-linux-intel64-rhel5.7/lib/python2.7/hashlib.py", line 139, in <module>

What do "near branches", "macro branches" and "taken speculatively" refer to?

I am looking at the list of branch events for Ivy Bridge and when finding their descriptions I come across the following terms:

"near branches"

"macro branches"

"taken speculative" 

What are the first two and what does the third mean? I thought the idea is to "predict" a branch, so I am unsure what the "speculative" means here?

The events in question are:

BR_MISP_RETIRED.NEAR_TAKEN

BR_MISP_RETIRED.ALL_BRANCHES

BR_INST_EXEC.TAKEN_INDIRECT_NEAR_CALL

Branching, instruction-miss and prefetcher event counters

Hi, would somebody be able to tell me which event counters I should add to a custom analysis for the following:

1) How successful the prefetcher is

2) L1 Instruction cache misses

3) Pipeline stalls regardless of cause

4) Direct Branch mispredictions

5) Indirect Branch mispredictions

Also, what is the difference between event names with and without the _PS, like:

BR_MISP_RETIRED.ALL_BRANCHES

BR_MISP_RETIRED.ALL_BRANCHES_PS

Thank you in advance,

 

hybrid application on the Xeon Phi

I am trying to execute a hybrid application (CP2K) on the Xeon Phi using MPI + OpenMP. I have the following environment set up: $ export OMP_NUM_THREADS=15 $ export I_MPI_PIN_PROCESSOR_LIST=$(seq -s "," 1 $OMP_NUM_THREADS 240) $ echo $I_MPI_PIN_PROCESSOR_LIST 1,16,31,46,61,76,91,106,121,136,151,166,181,196,211,226 $ mpirun -n $(expr 240 / $OMP_NUM_THREADS) cp2k.psmp.epcc H2O-64.inp When I run the "top" command only shows the 16 MPI processes and not any of the threads and says the Phi system is 6.2% user busy (16 / 240 * 100). It seems like the threads are not running.

icpc 15.0.0 internal errors when compiling C++11

I encountered two types of icpc internal errors when compiling some C++11 source files.  They look like:

$ icpc -std=c++11 -S lattice_numerical_constraint.ii
Internal error loop: assertion failed: find_seq_in_lookup_table: seq_number not found (shared/cfe/edgcpfe/il.c, line 3941)

compilation aborted for lattice_numerical_constraint.ii (code 4)

and

SIGABRT when throwing exception from code compiled with clang++ in subproject of Intel C++ project in OSX XCode 5 32 bit

I have a very simple 32 bit Xcode project setup running under OSX 10.9.5, Xcode 5.2

Main project - compiled by Intel C++ calling constructor of a class in a subproject. Subproject is a static library, compiled by llvm clang++.

In my main project (Intel C++) I have:

main.cpp:
#include "test.h"
int main(int argc, const char * argv[])
{
    test t;
    return 0;
}

In my subproject (LLVM Clang++) I have:

test.h:
class test {
public:
    test();
};

Subscribe to Parallel Computing