Article

Performance Tools for Software Developers - SSE generation and processor-specific optimizations continued

Can I combine the processor values and target more than one processor? How to generate optimized code for both Intel and AMD* architecture? Where can I find more information on processor-specific optimizations?
作者: 管理 最后更新时间: 2018/05/25 - 15:30
Article

Which applications are most likely to benefit from recompilation for Intel® Advanced Vector Extensions (Intel® AVX)?

Applications containing vectorizable, floating-point loops or calls to performance libraries are the most likely to see significant performance gains from rebuilding for the Intel® Advanced Vector Extensions (Intel® AVX)
作者: Martyn Corden (Intel) 最后更新时间: 2018/08/22 - 17:05
Article

Which applications are most likely to benefit from recompilation for Intel® Advanced Vector Extensions (Intel® AVX)?

Applications that spend considerable time in floating-point loops that can be vectorized are likely to benefit the most from the increased vector width of Intel® Advanced Vector Instructions (Intel® AVX).
作者: 管理 最后更新时间: 2018/05/25 - 15:30
博客

Parallel Universe Magazine #12: Advanced Vectorization

This blog contains additional content for the article "Advanced Vectorization" from Parallel Universe #12:

作者: 最后更新时间: 2019/07/03 - 20:08
Article
博客

Three Pieces of Advice for Code Modernization Success

What three code modernization techniques would I suggest to help a programmer improve the execution performance of her code? With too many specific things to choose from, these are three recommendations for any programmer anywhere and anytime.
作者: Clay B. (Blackbelt) 最后更新时间: 2018/12/12 - 18:08
File Wrapper

Parallel Universe Magazine - Issue 19, September 2014

作者: 管理 最后更新时间: 2018/12/12 - 18:08
File Wrapper

Parallel Universe Magazine - Issue 22, September 2015

作者: 管理 最后更新时间: 2018/12/12 - 18:08
File Wrapper

Parallel Universe Magazine - Issue 24, March 2016

作者: 管理 最后更新时间: 2018/12/12 - 18:08
Article

Intel® Math Kernel Library Improved Small Matrix Performance Using Just-in-Time (JIT) Code Generation for Matrix Multiplication (GEMM)

    The most commonly used and performance-critical Intel® Math Kernel Library (Intel® MKL) functions are the general matrix multiply (GEMM) functions.

作者: Gennady F. (Blackbelt) 最后更新时间: 2019/03/21 - 03:01