please note, that the slides of the conference's program have been uploaded to
On Wednesday 22 September 2010 17:53:12 Kenneth Lloyd wrote:
> Is that EuroMPI2010 ob1 paper publicly available? I get involved in various
> NUMA partitioning/architecting studies and it seems there is not a lot of
> discussion in this area.
> Ken Lloyd
> Kenneth A. Lloyd
> Watt Systems Technologies Inc.
> -----Original Message-----
> From: devel-bounces_at_[hidden] [mailto:devel-bounces_at_[hidden]] On
> Behalf Of Jeff Squyres Sent: Wednesday, September 22, 2010 6:00 AM
> To: Open MPI Developers
> Subject: Re: [OMPI devel] How to add a schedule algorithm to the pml
> Sorry for the delay in replying -- I was in Europe for the past two weeks;
> travel always makes me waaaay behind on my INBOX...
> On Sep 14, 2010, at 9:56 PM, å¼ æ¶ wrote:
> > I tried to add a schedule algorithm to the pml component ,ob1 etc. Poorly
> > I can only find a paper named "Open MPI: A Flexible High Performance
> > MPI" and some annotation in the source file. From them , I know ob1 has
> > implemented round-robin& weighted distribution algorithm. But after
> > tracking the MPI_Send(),I cann't figure out the location of these
> > implement ,let alone to add a new schedule algorithm. I have two
> > questions :
> > 1.The location of the schedule algorithm ?
> It's complicated -- I'd say that the PML is probably among the most
> complicated sections of Open MPI because it is the main "engine" that
> enforces the MPI point-to-point semantics. The algorithm is fairly well
> distribute throughout the PML source code. :-\
> > 2.There are five components :cm,crcpw ,csum ,ob1,V in the pml framework .
> > The function of these components?
> cm: this component drives the MTL point-to-point components. It is mainly
> a thin wrapper for network transports that provide their own MPI-like
> matching semantics. Hence, most of the MPI semantics are effectively done
> in the lower layer (i.e., in the MTL components and their dependent
> libraries). You probably won't be able to do much here, because such
> transports (MX, Portals, etc.) do most of their semantics in the network
> layer -- not in Open MPI. If you have a matching network layer, this is
> the PML that you probably use (MX, Portals, PSM).
> crcpw: this is a fork of the ob1 PML; it add some failover semantics.
> csum: this is also a fork of the ob1 PML; it adds checksumming semantics
> (so you can tell if the underlying transport had an error).
> v: this PML uses logging and replay to effect some level of fault
> tolerance. It's a distant fork of the ob1 PML, but has quite a few
> significant differences.
> ob1: this is the "main" PML that most users use (TCP, shared memory,
> OpenFabrics, etc.). It gangs together one or more BTLs to send/receive
> messages across individual network transports. Hence, it supports true
> multi-device/multi-rail algorithms. The BML (BTL multiplexing layer) is a
> thin management later that marshals all the BTLs in the process together
> -- it's mainly array handling, etc. The ob1 PML is the one that decides
> multi-rail/device splitting, etc. The INRIA folks just published a paper
> last week at Euro MPI about adjusting the ob1 scheduling algorithm to also
> take NUMA/NUNA/NUIOA effects into account, not just raw bandwidth
> Hope this helps!
Dr.-Ing. Rainer Keller http://www.hlrs.de/people/keller
HLRS Tel: ++49 (0)711-685 6 5858
Nobelstrasse 19 Fax: ++49 (0)711-685 6 5832
70550 Stuttgart email: keller_at_[hidden]