> > You'll want to look at orte/mca/rmaps/rank_file/rmaps_rank_file.c
> > - the bit map is now computed in mpirun and then sent to the daemons
> Actually, I'm getting lost in this code. Anyhow, I don't think
> the problem is related to Solaris. I think it's also on Linux.
> E.g., I can reproduce the problem with 1.9a1r28035 on Linux using
> GCC compilers.
> Siegmar: can you confirm this is a problem also on Linux? E.g.,
> with OMPI 1.9, on one of your Linux nodes (linpc0?) try
> % cat myrankfile
> rank 0=linpc0 slot=0:1
> % mpirun --report-bindings --rankfile myrankfile numactl --show
> For me, the binding I get is not 0:1 but 0,1.
I get the following outputs for openmpi-1.6.4rc4 (without your patch)
and openmpi-1.9 (both compiled with Sun C 5.12).
linpc1 rankfiles 96 cat rf_1_linux
rank 0=linpc1 slot=0:1
linpc1 rankfiles 97 mpirun --report-bindings --rankfile rf_1_linux \
[linpc1:15882] MCW rank 0 bound to socket 0[core 1]:
[. B][. .] (slot list 0:1)
preferred node: 0
membind: 0 1
linpc1 rankfiles 98 ompi_info | grep "MPI:"
Open MPI: 1.6.4rc4r28022
linpc1 rankfiles 96 mpirun --report-bindings --rankfile rf_1_linux \
[linpc1:16061] MCW rank 0 bound to socket 0[core 0[hwt 0]],
socket 0[core 1[hwt 0]]: [B/B][./.]
preferred node: current
physcpubind: 0 1
membind: 0 1
linpc1 rankfiles 97 ompi_info | grep "MPI:"
Open MPI: 1.9a1r28035
Today I will build the current versions of openmpi-1.6.4rc4 and
openmpi-1.9, so that I can test a rankfile with two machines