This two day webinar series introduces you to the world of multicore and manycore computing with Intel® Xeon® processors and Intel® Xeon Phi™ coprocessors. Expert technical teams at Intel discuss development tools, programming models, vectorization, and execution models that will get your development efforts powered up to get the best out of your applications and platforms.
I am attempting to use the nanosleep function (defined here: http://linux.die.net/man/2/nanosleep) on the Xeon Phi. It is very similar to the standard (and by standard I just mean widely-used, I know there is no true standard) sleep function, but it is higher-resolution, a requirement for our software.
Intel(r) IPP offers two ways to try out Android* support:
1. Intel(r) IPP Preview for Android* available exclusively from Beacon Mountain
- This is a subset of Intel(r) IPP image processing and computer vision functions that is freely available for download, use, and production release for Android platforms. http://software.intel.com/en-us/vcsource/tools/beaconmountain
2. Intel(r) IPP for Android*
One can use API __Offload_report ( int ) or the environmental variable OFFLOAD_REPORT to get the timing information ( and the amount of data transferred ) for all the offloaded regions. And the profiling information is output to the stdout.
But for profiling tool developers, the following features may be desired:
Is it possible to provide a profiling API that can be used to timing a specific offloaded region , instread of all offloaded regions?
I'm having some trouble getting Automatic Offload to work with the MKL dgetrf & dgetri routines on our server with two Phi cards. dgemm routines in this code work just fine. Here's the build code -
icpc -c -fpic -shared -std=c++11 -O3 -xHost -ip -ipo3 -parallel -funroll-loops -fno-alias -fno-fnalias -fargument-noalias -mkl -I include/ -I ~/Documents/Boost/boost_1_53_0/ src/PRH.cpp -o src/obj/PRH.o
Here's the OFFLOAD_REPORT generated when the code runs -