Run "ibstat" on each host:
1. Make sure the adapters are alive and active.
2. Look at the Link Layer settings for host w34. Does it match host w4's?
On Fri, May 9, 2014 at 1:18 PM, Tim Miller <btamiller_at_[hidden]> wrote:
> Hi All,
> We're using OpenMPI 1.7.3 with Mellanox ConnectX InfiniBand adapters, and
> periodically our jobs abort at start-up with the following error:
> Open MPI detected two different OpenFabrics transport types in the same
> Infiniband network.
> Such mixed network trasport configuration is not supported by Open MPI.
> Local host: w4
> Local adapter: mlx4_0 (vendor 0x2c9, part ID 26428)
> Local transport type: MCA_BTL_OPENIB_TRANSPORT_IB
> Remote host: w34
> Remote Adapter: (vendor 0x2c9, part ID 26428)
> Remote transport type: MCA_BTL_OPENIB_TRANSPORT_UNKNOWN
> I've done a bit of googling and not found very much. We do not see this
> issue when we run with MVAPICH2 on the same sets of nodes.
> Any advice or thoughts would be very welcome, as I am stumped by what
> causes this. The nodes are all running Scientific Linux 6 with Mellanox
> drivers installed via the SL-provided RPMs.
> users mailing list