Article

OpenMP* Loop Scheduling

Compiler Methodology for Intel® MIC Architecture

Authored by admin Last updated on 09/30/2019 - 17:28
Article

OpenMP Loop Collapse Directive

Compiler Methodology for Intel® MIC Architecture

Authored by admin Last updated on 09/30/2019 - 17:30
Article

OpenMP* Thread Affinity Control

Compiler Methodology for Intel® MIC Architecture

Authored by admin Last updated on 09/30/2019 - 17:28
Article

OpenMP Related Tips

Compiler Methodology for Intel® MIC Architec

Authored by AmandaS (Intel) Last updated on 09/30/2019 - 17:28
File Wrapper

Parallel Universe Magazine - Issue 20, February 2015

Authored by admin Last updated on 12/12/2018 - 18:08
Article

Efficient Parallelization

This article is part of the Intel® Modern Code Developer Community documentation which supports developers in leveraging application performance in code through a systematic step-by-step optimization framework methodology. This article addresses: Thread level parallelization.
Authored by Ronald W Green (Blackbelt) Last updated on 09/30/2019 - 17:28
Article

New Features for Intel® MIC Architecture in the Intel Compiler

The list below summarizes new features and changes specific to programming for Intel® MIC Architecture with Intel Compiler 15.0:

Authored by AmandaS (Intel) Last updated on 10/15/2019 - 16:40
Article

Scheduling for 1-4 Threads Per Core Using Compiler Option -qopt-threads-per-core

Compiler Methodology for Intel® MIC Architecture

Authored by admin Last updated on 09/30/2019 - 17:30
Article

Loop Modifications to Enhance Data-Parallel Performance

When confronted with nested loops, the granularity of the computations that are assigned to threads will directly affect performance. Loop transformations such as splitting and merging nested loops can make parallelization easier and more productive.
Authored by admin Last updated on 07/05/2019 - 14:47
Article

Granularity and Parallel Performance

One key to attaining good parallel performance is choosing the right granularity for the application. Granularity is the amount of real work in the parallel task. If granularity is too fine, then performance can suffer from communication overhead.
Authored by admin Last updated on 07/05/2019 - 19:52