过滤器

Article

Intel® Fortran Compiler for Linux* - Are the libraries thread safe?

Are the Intel Fortran run-time libraries thread safe?
作者: 管理 最后更新时间: 2019/07/04 - 10:00
Article

Requirements for Vectorizable Loops

Vectorization is one of many optimizations that are enabled by default in the latest Intel compilers. In order to be vectorized, loops must obey certain conditions, listed below. Some additional ways to help the compiler to vectorize loops are described.
作者: Martyn Corden (Intel) 最后更新时间: 2019/03/27 - 14:36
Article

The Three Stages of Preparation for Optimizing Parallel Software

Improving software performance on parallel software requires a structured approach that makes good use of development resources, obtaining good results quickly.

作者: aaron-tersteeg (Intel) 最后更新时间: 2019/07/05 - 10:15
Article

Determining Root Cause of Segmentation Faults SIGSEGV or SIGBUS errors

SIGSEGV on Linux and SIGBUS on MacOS Root Causes
作者: 管理 最后更新时间: 2018/12/26 - 14:09
Article

Intel® MKL Threaded 1D FFTs

This document describes the cases for which the Intel MKL 10.2 and later 1D complex-to-complex FFTs are threaded.
作者: 最后更新时间: 2019/03/27 - 10:00
Article

Threading Fortran Applications for Parallel Performance on Multi-Core Systems

Advice and background information is given on typical issues that may arise when threading an application using the Intel Fortran Compiler and other software tools, whether using OpenMP, automatic parallelization or threaded libraries.
作者: Martyn Corden (Intel) 最后更新时间: 2018/12/12 - 18:00
Article

Loop Modifications to Enhance Data-Parallel Performance

When confronted with nested loops, the granularity of the computations that are assigned to threads will directly affect performance. Loop transformations such as splitting and merging nested loops can make parallelization easier and more productive.
作者: 管理 最后更新时间: 2019/07/05 - 14:47
Article

Granularity and Parallel Performance

One key to attaining good parallel performance is choosing the right granularity for the application. Granularity is the amount of real work in the parallel task. If granularity is too fine, then performance can suffer from communication overhead.
作者: 管理 最后更新时间: 2019/07/05 - 19:52
Article

Expose Parallelism by Avoiding or Removing Artificial Dependencies

Many applications and algorithms contain serial optimizations that inadvertently introduce data dependencies and inhibit parallelism. One can often remove such dependences through simple transforms, or even avoid them altogether through.
作者: 管理 最后更新时间: 2019/07/05 - 19:49