To answer the original questions, Open MPI will look at taking advantage of the RDMA CUDA when it is available.  Obviously, work needs to be done to figure out the best way to integrate into the library.  Much like there are a variety of protocols under the hood to support host transfer of data via IB, we will have to see what works  best for transferring GPU buffers.


It is unclear how this will affect the send/receive latency. 


Lastly, the support will be for Kepler –class Quadro and Tesla devices. 





From: [] On Behalf Of Durga Choudhury
Sent: Tuesday, August 14, 2012 4:46 PM
To: Open MPI Users
Subject: Re: [OMPI users] RDMA GPUDirect CUDA...


Dear OpenMPI developers

I'd like to add my 2 cents that this would be a very desirable feature enhancement for me as well (and perhaps others).

Best regards

On Tue, Aug 14, 2012 at 4:29 PM, Zbigniew Koza <> wrote:


I've just found this information on  nVidia's plans regarding enhanced support for MPI in their CUDA toolkit:

The idea that two GPUs can talk to each other via network cards without CPU as a middleman looks very promising.
This technology is supposed to be revealed and released in September.

My questions:

1. Will OpenMPI include   RDMA support in its CUDA interface?
2. Any idea how much can this technology reduce the CUDA Send/Recv latency?
3. Any idea whether this technology will be available for Fermi-class Tesla devices or only for Keplers?


Z Koza

users mailing list


This email message is for the sole use of the intended recipient(s) and may contain confidential information.  Any unauthorized review, use, disclosure or distribution is prohibited.  If you are not the intended recipient, please contact the sender by reply email and destroy all copies of the original message.