Computação de cluster

How to use Intel® Cluster Checker v3 when the cluster head node has an Intel® Xeon Phi coprocessor installed

In case your cluster uses a head node with one or more Intel® Xeon PhiTM coprocessors a couple more step have to be taken in order to collect all information.

First of all execute the clck-collect command using your configuration file and node list.

clck-collect -a -c <your configuration file> -f <your node list>

Then create a second node list containing just the name of the head node, without any additional "role:" tag.

Now you can collect the additional information by executing

  • Parceiros
  • Linux*
  • Servidor
  • Intel® Cluster Checker
  • Intel® Cluster Ready
  • Computação de cluster
  • Intel MPI® Library 5.0 Update 3 (Build 049) Readme

    The Intel® MPI Library is a high-performance interconnect-independent multi-fabric library implementation of the industry-standard Message Passing Interface, v3.0 (MPI-3.0) specification. This package is for MPI users who develop on and build for Intel® 64 architectures on Linux*, as well as customers running on the Intel® Xeon Phi™ coprocessor on Linux*. You must have a valid license to download, install, and use this product.

  • Linux*
  • C/C++
  • Fortran
  • Biblioteca MPI Intel®
  • Interface de transferência de mensagens
  • Computação de cluster
  • iconv issue

    hi all,


    I'm trying to build something for the Phi that depends on iconv; the library routines are present , but the following application fails when run on the Phi:

    #include <stdlib.h>
    #include <iconv.h>
    int main () {
      iconv_t cd;
      cd = iconv_open("latin1","UTF-8");
      if(cd == (iconv_t)(-1)) exit(1);

    if I build this using "icc -o iconv_test iconv_test.c" and run it on the host it return no error (exit code 0).

    No Cost Options for Intel Parallel Studio XE, Support yourself, Royalty-Free

    Intel® Parallel Studio XE is a very popular product from Intel that includes the Intel® Compilers, Intel® Performance Libraries, tools for analysis, debugging and tuning, tools for MPI and the Intel® MPI Library. Did you know that some of these are available for free?

    Here is a guide to “what is available free” from the Intel Parallel Studio XE suites.

    Intel® Parallel Studio XE 2016: High Performance for HPC Applications and Big Data Analytics

    Intel® Parallel Studio XE 2016, launched on August 25, 2015, is the latest installment in our developer toolkit for high performance computing (HPC) and technical computing applications. This suite of compilers, libraries, debugging facilities, and analysis tools, targets Intel® architecture, including support for the latest Intel® Xeon® processors (codenamed Skylake) and Intel® Xeon Phi™ processors (codenamed Knights Landing).

    Need help making sense of NBC performance (MPI3)

    Hello everyone,

    I am fairly new to parallel computing, but am working on a certain legacy code that uses real-space domain decomposition for electronic structure calculations. I have spent a while modernizing the main computational kernel to hybrid MPI+openMP and upgraded the communication pattern to use nonblocking neighborhood alltoallv for the halo exchange and a nonblocking allreduce for the other communication in the kernel. I have now started to focus on "communication hiding", so that the calculations and communication happen alongside each other.

    Can each thread on Xeon Phi be given private data areas in the offload model


    I want to calculate a  Jacobian matrix, which is a sum of 960 (to be simple) 3x3 matrices  by distributing the calculations of these 3x3 matrices to a Xeon Phi card. The calculation of the 3x3 matrices uses a third party library whose subroutines use an interger vector not only for the storage of parameter values but also to write and read intermidiate results. It is therefore necessary for each task to have this integer vector protected from other tasks. Can this be obtained on the physical core level or even for each thread (each Xeon Phi has 60x4=240 threads. 

    mpirun with bad hostname hangs with [ssh] <defunct> until Enter is pressed

    We have been experiencing hangs with our MPI-based application and our investigation led us to observing the following behaviour of mpirun:

    mpirun -n 1 -host <good_hostname> hostname works as expected

    mpirun -n 1 -host <bad_hostname> hostname hangs, during which ps shows: 

    Assine o Computação de cluster