Open MPI logo

Open MPI Development Mailing List Archives

  |   Home   |   Support   |   FAQ   |   all Development mailing list

Subject: Re: [OMPI devel] poor btl sm latency
From: Jeffrey Squyres (jsquyres_at_[hidden])
Date: 2012-03-15 12:10:33

On Mar 15, 2012, at 8:06 AM, Matthias Jurenz wrote:

> We made a big step forward today!
> The used Kernel has a bug regarding to the shared L1 instruction cache in AMD
> Bulldozer processors:
> See
> and
> Until the Kernel is patched we disable the address-space layout randomization
> (ASLR) as described in the above PDF:
> $ sudo /sbin/sysctl -w kernel.randomize_va_space=0
> Therewith, NetPIPE results in ~0.5us latency when binding the processes for
> L2/L1I cache sharing (i.e. -bind-to-core).

This is good! I love it when the bug is not our fault. :-)

> However, when binding the processes for exclusive L2/L1I caches (i.e. -cpus-
> per-proc 2) we still get ~1.1us latency. I don't think that the upcoming
> kernel patch will help for this kind of process binding...

Does this kind of thing happen with Platform MPI, too? I.e., is this another kernel issue, or an OMPI-specific issue?

Jeff Squyres
For corporate legal information go to: