Article
Article

Code Sample: Exploring MPI for Python* on Intel® Xeon Phi™ Processor

Learn how to write an MPI program in Python*, and take advantage of Intel® multicore architectures using OpenMP threads and Intel® AVX512 instructions.
Authored by Nguyen, Loc Q (Intel) Last updated on 07/06/2019 - 16:30
Article

Optimization Techniques for the Intel® MIC Architecture: Part 1 of 3

Part one of this three-part series focuses on thread parallelism and race conditions, and discusses using mutexes in OpenMP* to resolve race conditions.
Authored by Mike P. (Intel) Last updated on 03/21/2019 - 12:00
Article

Process and Thread Affinity for Intel® Xeon Phi™ Processors

The Intel® MPI Library and OpenMP* runtime libraries can create affinities between processes or threads, and hardware resources. This affinity keeps an MPI process or OpenMP thread from migrating to a different hardware resource, which can have a dramatic effect on the execution speed of a program.
Authored by Gregg S. (Intel) Last updated on 03/21/2019 - 12:00
Article

著作 - High Performance Parallelism Pearls

A look into the contents of the two "Pearls" books, edited by James Reinders and Jim Jeffers. These books contain a collection of examples of code modernization.
Authored by Mike P. (Intel) Last updated on 03/21/2019 - 12:00
Article

面向英特尔® 至强融核™ 处理器(代号“Knights Landing”)的开发人员访问计划

Intel is bringing to market, in anticipation of general availability of the Intel® Xeon Phi™ Processor (codenamed Knights Landing), the Developer Access Program (DAP). DAP is an early access program for developers worldwide to purchase an Intel Xeon Phi Processor based system.
Authored by Mike P. (Intel) Last updated on 03/21/2019 - 12:00
Article

Books - High Performance Parallelism Pearls

A look into the contents of the two "Pearls" books, edited by James Reinders and Jim Jeffers. These books contain a collection of examples of code modernization.
Authored by Mike P. (Intel) Last updated on 03/21/2019 - 12:00
Article

Hybrid Parallelism: Parallel Distributed Memory and Shared Memory Computing

There are two principal methods of parallel computing: distributed memory computing and shared memory computing. As more processor cores are dedicated to large clusters solving scientific and engineering problems, hybrid programming techniques combining the best of distributed and shared memory programs are becoming more popular.
Authored by David M. Last updated on 07/12/2019 - 08:31
Article

面向英特尔® 至强融核™ 处理器的 Offload over Fabric教程

This tutorial shows how to install Offload over Fabric (OoF) software on 2nd generation Intel® Xeon Phi™ processor, configure the hardware, test the basic configuration, and enable OoF
Authored by Nguyen, Loc Q (Intel) Last updated on 03/21/2019 - 12:00
Article

Choosing the right threading framework

This is the second article in a series of articles about High Performance Computing with the Intel Xeon Phi.

Authored by Last updated on 07/06/2019 - 16:30