To answer the original questions, Open MPI will look at taking advantage of the RDMA CUDA when it is available. Obviously, work needs to be done to figure out the best way to integrate into the library. Much like there are a variety of protocols under the hood to support host transfer of data via IB, we will have to see what works best for transferring GPU buffers.
It is unclear how this will affect the send/receive latency.
Lastly, the support will be for Kepler –class Quadro and Tesla devices.
From: email@example.com [mailto:firstname.lastname@example.org] On Behalf Of Durga Choudhury
Sent: Tuesday, August 14, 2012 4:46 PM
To: Open MPI Users
Subject: Re: [OMPI users] RDMA GPUDirect CUDA...
Dear OpenMPI developers
I'd like to add my 2 cents that this would be a very desirable feature enhancement for me as well (and perhaps others).
On Tue, Aug 14, 2012 at 4:29 PM, Zbigniew Koza <email@example.com> wrote:
I've just found this information on nVidia's plans regarding enhanced support for MPI in their CUDA toolkit:
The idea that two GPUs can talk to each other via network cards without CPU as a middleman looks very promising.
This technology is supposed to be revealed and released in September.
1. Will OpenMPI include RDMA support in its CUDA interface?
2. Any idea how much can this technology reduce the CUDA Send/Recv latency?
3. Any idea whether this technology will be available for Fermi-class Tesla devices or only for Keplers?
users mailing list
This email message is for the sole use of the intended recipient(s) and may
contain confidential information. Any unauthorized review, use, disclosure
or distribution is prohibited. If you are not the intended recipient,
please contact the sender by reply email and destroy all copies of the original