Blog post

Hybrid MPI and OpenMP* Model

In the High Performance Computing (HPC) area, parallel computing techniques such as MPI, OpenMP*, one-sided communications, shmem, and Fortran coarray are widely utilized. This blog is part of a series that will introduce the use of these techniques, especially how to use them on the Intel® Xeon Phi™ coprocessor. This first blog discusses the main usage of the hybrid MPI/OpenMP model.
Authored by Nguyen, Loc Q (Intel) Last updated on 07/06/2019 - 17:10
Article

Hybrid applications: Intel MPI Library and OpenMP*

Tips and tricks on how to get the optimal performance settings for your mixed Intel MPI/OpenMP applications.
Authored by Gergana S. (Blackbelt) Last updated on 07/06/2019 - 19:20
Article

Books - Message Passing Interface (MPI)

This article looks at several books that introduce developers to the topics of Message Passing Interface (MPI), parallel programming, and OpenMP*.
Authored by Mike P. (Intel) Last updated on 12/12/2018 - 18:00
Article

Process and Thread Affinity for Intel® Xeon Phi™ Processors

The Intel® MPI Library and OpenMP* runtime libraries can create affinities between processes or threads, and hardware resources. This affinity keeps an MPI process or OpenMP thread from migrating to a different hardware resource, which can have a dramatic effect on the execution speed of a program.
Authored by Gregg S. (Intel) Last updated on 07/29/2019 - 08:05
Article

Scale-Up Implementation of a Transportation Network Using Ant Colony Optimization (ACO)

In this article an OpenMP* based implementation of the Ant Colony Optimization algorithm was analyzed for bottlenecks with Intel® VTune™ Amplifier XE 2016 together with improvements using hybrid MPI-OpenMP and Intel® Threading Building Blocks were introduced to achieve efficient scaling across a four-socket Intel® Xeon® processor E7-8890 v4 processor-based system.
Authored by Sunny G. (Intel) Last updated on 07/05/2019 - 19:10
File Wrapper

Parallel Universe Magazine - Issue 24, March 2016

Authored by admin Last updated on 12/12/2018 - 18:08
Article

Hybrid Parallelism: Parallel Distributed Memory and Shared Memory Computing

There are two principal methods of parallel computing: distributed memory computing and shared memory computing. As more processor cores are dedicated to large clusters solving scientific and engineering problems, hybrid programming techniques combining the best of distributed and shared memory programs are becoming more popular.
Authored by David M. Last updated on 07/12/2019 - 08:31
Article

An Introduction to MPI-3 Shared Memory Programming

In this article, we present a tutorial on how to start using MPI SHM on multinode systems using Intel® Xeon® and Intel® Xeon Phi™ processors. The article uses a 1-D ring application as an example and includes code snippets to describe how to transform common MPI send/receive patterns to utilize the MPI SHM interface. The MPI functions that are necessary for internode and intranode communications...
Authored by Last updated on 07/27/2018 - 08:58
Article

Combining Linux* Message Passing and Threading in High Performance Computing

An article addressing thread and task parallelism. This article can be used to optimize framework methodology. Written by Andrew Binstock--Principal Analyst at Pacific Data Works LLC and lead author of "Practical Algorithms for Programmers."
Authored by Last updated on 07/06/2019 - 16:22
Article

Classical Molecular Dynamics Simulations with LAMMPS Optimized for Knights Landing

LAMMPS is an open-source software package that simulates classical molecular dynamics. As it supports many energy models and simulation options, its versatility has made it a popular choice. It was first developed at Sandia National Laboratories to use large-scale parallel computation.
Authored by WILLIAM B. (Intel) Last updated on 03/21/2019 - 12:00