Parallel Computing

Intel® Xeon Phi™ coprocessor hardware and software architecture introduction

This two day webinar series introduces you to the world of multicore and manycore computing with Intel® Xeon® processors and Intel® Xeon Phi™ coprocessors. Expert technical teams at Intel discuss development tools, programming models, vectorization, and execution models that will get your development efforts powered up to get the best out of your applications and platforms.

  • Developers
  • Server
  • Intel Xeon Phi
  • Intel Xeon Phi Coprocessor
  • offload
  • Optimization
  • Parallel Computing
  • Vectorization
  • Intel(r) Integrated Performance Primitives Android* Support

    Intel(r) IPP offers two ways to try out Android* support:

    1. Intel(r) IPP Preview for Android* available exclusively from Beacon Mountain

    2. Intel(r) IPP for Android*

    Any profiling API for tool developers?

    Hi,

    One can use API __Offload_report ( int )  or the environmental variable OFFLOAD_REPORT to get the timing information ( and the amount of data transferred ) for all the offloaded regions. And the profiling information is output to the stdout.

    But for profiling tool developers, the following features may be desired:

    Is it possible to provide a profiling API that can be used to timing a specific offloaded region , instread of all offloaded regions?

    Automatic Offload not working for dgetrf, dgetri

    I'm having some trouble getting Automatic Offload to work with the MKL dgetrf & dgetri routines on our server with two Phi cards. dgemm routines in this code work just fine. Here's the build code -

    icpc -c -fpic -shared  -std=c++11 -O3 -xHost -ip -ipo3 -parallel -funroll-loops -fno-alias -fno-fnalias -fargument-noalias -mkl -I include/ -I ~/Documents/Boost/boost_1_53_0/ src/PRH.cpp -o src/obj/PRH.o

    Here's the OFFLOAD_REPORT generated when the code runs -

    Subscribe to Parallel Computing