This web mail archive is frozen.
This page is part of a frozen web archive of this mailing list.
You can still navigate around this archive, but know that no new mails
have been added to it since July of 2016.
Click here to be taken to the new web archives of this list; it includes all the mails that are in this frozen archive plus all new mails that have been sent to the list since it was migrated to the new archives.
For those not following the user list, this request was generated today:
>>>> Absolutely, these are useful time and time again so should be part of
>>>> the API and hence stable. Care to mention what they are and I'll add it
>>>> to my note as something to change when upgrading to 1.3 (we are looking
>>>> at testing a snapshot in the near future).
>>> OMPI_COMM_WORLD_SIZE #procs in the job
>>> OMPI_COMM_WORLD_LOCAL_SIZE #procs in this job that are sharing the node
>>> OMPI_UNIVERSE_SIZE total #slots allocated to this user
>>> (across all nodes)
>>> OMPI_COMM_WORLD_RANK proc's rank
>>> OMPI_COMM_WORLD_LOCAL_RANK local rank on node - lowest rank'd proc on
>>> the node is given local_rank=0
>>> If there are others that would be useful, now is definitely the time to
>>> speak up!
>> The only other one I'd like to see is some kind of global identifier for
>> the job but as far as I can see I don't believe that openmpi has such a
> Not really - of course, many environments have a jobid they assign at time
> of allocation. We could create a unified identifier from that to ensure a
> consistent name was always available, but the problem would be that not all
> environments provide it (e.g., rsh). To guarantee that the variable would
> always be there, we would have to make something up in those cases.
I could easily create such an envar, even for non-managed environments. The
plan would be to use the RM-provided jobid where one was available, and to
use the mpirun jobid where not.
My thought was to call it OMPI_JOB_ID, unless someone has another
Any objection to my doing so, and/or suggestions on alternative