English
 
Help Privacy Policy Disclaimer
  Advanced SearchBrowse

Item

ITEM ACTIONSEXPORT
 
 
DownloadE-Mail
  Overlapping communication and computation using the Intel MPI library's asynchronous progress control

Ohlmann, S., Baruffa, F., & Rampp, M. (2020). Overlapping communication and computation using the Intel MPI library's asynchronous progress control. Talk presented at IXPUG Annual Meeting 2020 (IXPUG: Intel eXtreme Performance Users Group). Virtually Hosted by TACC. 2020-10-13 - 2020-10-16.

Item is

Files

show Files
hide Files
:
Overlapping communication and computation using the Intel MPI librarys asynchronous progress control.pdf (Any fulltext), 6MB
 
File Permalink:
-
Name:
Overlapping communication and computation using the Intel MPI librarys asynchronous progress control.pdf
Description:
-
OA-Status:
Visibility:
Private
MIME-Type / Checksum:
application/pdf
Technical Metadata:
Copyright Date:
-
Copyright Info:
-
License:
-

Locators

show

Creators

show
hide
 Creators:
Ohlmann, Sebastian1, Author           
Baruffa, Fabio, Author
Rampp, Markus1, Author           
Affiliations:
1Max Planck Computing and Data Facility, Max Planck Society, ou_2364734              

Content

show
hide
Free keywords: Intel MPI, Asynchronous progress, Communication-computation overlap, Stencil code, Domain decomposition, Electronic structure, Octopus
 Abstract: When scaling HPC applications to large-scale systems, the
time spent in communication often becomes a bottleneck. A well-known
technique to tackle this problem is overlapping communication and
computation to hide communication time. In MPI codes, however, using
non-blocking functions is not enough -- the progress of the
communication needs to be triggered explicitly, either by the
application code or by special features of the MPI library. In this
talk, we explore overlapping communication and computation using the
asynchronous progress control feature of the Intel(r) MPI library by
applying it to stencil codes with a domain decomposition. With this
feature, the MPI library transparently handles the progress of
non-blocking MPI communication, removing the need for an explicit
control in the application. First, we introduce the asynchronous
progress control of the Intel(r) MPI library and how it can be used to
improve performance and scalability of a simple domain-decomposition
code. Second, we show how a real-world application can benefit from
this feature: the electronic structure code Octopus that uses finite
difference stencils to solve the time-dependent DFT equations.
Moreover, we try to generalize the conditions under which other
stencil codes with a domain decomposition can benefit as well. All
tests have been run on Cobra, the current flagship system of the Max
Planck Society with about 3400 nodes (having 2 Intel(r) Skylake 6148
Gold sockets each) and an Omnipath interconnect, hosted at the Max
Planck Computing and Data Facility (MPCDF).

Details

show
hide
Language(s): eng - English
 Dates: 2020-10-13
 Publication Status: Published online
 Pages: -
 Publishing info: -
 Table of Contents: -
 Rev. Type: -
 Identifiers: -
 Degree: -

Event

show
hide
Title: IXPUG Annual Meeting 2020 (IXPUG: Intel eXtreme Performance Users Group)
Place of Event: Virtually Hosted by TACC
Start-/End Date: 2020-10-13 - 2020-10-16

Legal Case

show

Project information

show

Source

show