2005 | OriginalPaper | Chapter
Supporting MPI-2 One Sided Communication on Multi-rail InfiniBand Clusters: Design Challenges and Performance Benefits
Authors : Abhinav Vishnu, Gopal Santhanaraman, Wei Huang, Hyun-Wook Jin, Dhabaleswar K. Panda
Published in: High Performance Computing – HiPC 2005
Publisher: Springer Berlin Heidelberg
Activate our intelligent search to find suitable subject content or patents.
Select sections of text to find matching patents with Artificial Intelligence. powered by
Select sections of text to find additional relevant content using AI-assisted search. powered by
In cluster computing, InfiniBand has emerged as a popular high performance interconnect with MPI as the
de facto
programming model. However, even with InfiniBand, bandwidth can become a bottleneck for clusters executing communication intensive applications. Multi-rail cluster configurations with MPI-1 are being proposed to alleviate this problem. Recently, MPI-2 with support for one-sided communication is gaining significance. In this paper, we take the challenge of designing high performance MPI-2
one-sided communication
on multi-rail InfiniBand clusters. We propose a unified MPI-2 design for different configurations of multi-rail networks (
multiple ports, multiple HCAs
and
combinations
). We present various issues associated with one-sided communication such as
multiple synchronization messages, scheduling of RDMA (Read, Write) operations, ordering relaxation
and discuss their implications on our design. Our performance results show that multi-rail networks can significantly improve MPI-2 one-sided communication performance. Using PCI-Express with two-ports, we can achieve a peak
MPI_Put
bidirectional bandwidth of 2620 Million Bytes/s, compared to 1910 MB/s for single-rail implementation. For PCI-X with two HCAs, we can almost double the throughput and reduce the latency to half for large messages.