Towards Efficient HPC: Exploring Overlap Strategies Using MPI Non-Blocking Communication

As high-performance computing (HPC) platforms continue to scale up, communication costs have become a critical bottleneck affecting overall application performance. An effective strategy to overcome this limitation is to overlap communication with computation. The Message Passing Interface (MPI), as...

Full description

Saved in:
Bibliographic Details
Main Authors: Yuntian Zheng, Jianping Wu
Format: Article
Language:English
Published: MDPI AG 2025-06-01
Series:Mathematics
Subjects:
Online Access:https://www.mdpi.com/2227-7390/13/11/1848
Tags: Add Tag
No Tags, Be the first to tag this record!
_version_ 1850129458665619456
author Yuntian Zheng
Jianping Wu
author_facet Yuntian Zheng
Jianping Wu
author_sort Yuntian Zheng
collection DOAJ
description As high-performance computing (HPC) platforms continue to scale up, communication costs have become a critical bottleneck affecting overall application performance. An effective strategy to overcome this limitation is to overlap communication with computation. The Message Passing Interface (MPI), as the de facto standard for communication in HPC, provides non-blocking communication primitives that make such overlapping feasible. By enabling asynchronous communication, non-blocking operations reduce idle time of cores caused by data transfer delays, thereby improving resource utilization. Overlapping communication with computation is particularly important for enhancing the performance of large-scale scientific applications, such as numerical simulations, climate modeling, and other data-intensive tasks. However, achieving efficient overlapping is non-trivial and depends not only on advances in hardware technologies such as Remote Direct Memory Access (RDMA), but also on well-designed and optimized MPI implementations. This paper presents a comprehensive survey on the principles of MPI non-blocking communication, the core techniques for achieving computation–communication overlap, and some representative applications in scientific computing. Alongside the survey, we include a preliminary experimental study evaluating the effectiveness of asynchronous progress mechanism on modern HPC platforms to support the development of parallel programs for HPC researchers and practitioners.
format Article
id doaj-art-6f004068c82e4a01a60b7c1b3631f6bc
institution OA Journals
issn 2227-7390
language English
publishDate 2025-06-01
publisher MDPI AG
record_format Article
series Mathematics
spelling doaj-art-6f004068c82e4a01a60b7c1b3631f6bc2025-08-20T02:32:57ZengMDPI AGMathematics2227-73902025-06-011311184810.3390/math13111848Towards Efficient HPC: Exploring Overlap Strategies Using MPI Non-Blocking CommunicationYuntian Zheng0Jianping Wu1College of Meteorology and Oceanography, National University of Defense Technology, Changsha 410073, ChinaCollege of Meteorology and Oceanography, National University of Defense Technology, Changsha 410073, ChinaAs high-performance computing (HPC) platforms continue to scale up, communication costs have become a critical bottleneck affecting overall application performance. An effective strategy to overcome this limitation is to overlap communication with computation. The Message Passing Interface (MPI), as the de facto standard for communication in HPC, provides non-blocking communication primitives that make such overlapping feasible. By enabling asynchronous communication, non-blocking operations reduce idle time of cores caused by data transfer delays, thereby improving resource utilization. Overlapping communication with computation is particularly important for enhancing the performance of large-scale scientific applications, such as numerical simulations, climate modeling, and other data-intensive tasks. However, achieving efficient overlapping is non-trivial and depends not only on advances in hardware technologies such as Remote Direct Memory Access (RDMA), but also on well-designed and optimized MPI implementations. This paper presents a comprehensive survey on the principles of MPI non-blocking communication, the core techniques for achieving computation–communication overlap, and some representative applications in scientific computing. Alongside the survey, we include a preliminary experimental study evaluating the effectiveness of asynchronous progress mechanism on modern HPC platforms to support the development of parallel programs for HPC researchers and practitioners.https://www.mdpi.com/2227-7390/13/11/1848HPCMPInon-blocking communicationcomputation–communication overlapparallel computingasynchronous communication
spellingShingle Yuntian Zheng
Jianping Wu
Towards Efficient HPC: Exploring Overlap Strategies Using MPI Non-Blocking Communication
Mathematics
HPC
MPI
non-blocking communication
computation–communication overlap
parallel computing
asynchronous communication
title Towards Efficient HPC: Exploring Overlap Strategies Using MPI Non-Blocking Communication
title_full Towards Efficient HPC: Exploring Overlap Strategies Using MPI Non-Blocking Communication
title_fullStr Towards Efficient HPC: Exploring Overlap Strategies Using MPI Non-Blocking Communication
title_full_unstemmed Towards Efficient HPC: Exploring Overlap Strategies Using MPI Non-Blocking Communication
title_short Towards Efficient HPC: Exploring Overlap Strategies Using MPI Non-Blocking Communication
title_sort towards efficient hpc exploring overlap strategies using mpi non blocking communication
topic HPC
MPI
non-blocking communication
computation–communication overlap
parallel computing
asynchronous communication
url https://www.mdpi.com/2227-7390/13/11/1848
work_keys_str_mv AT yuntianzheng towardsefficienthpcexploringoverlapstrategiesusingmpinonblockingcommunication
AT jianpingwu towardsefficienthpcexploringoverlapstrategiesusingmpinonblockingcommunication