>>> You sound like our vendors, "what is your app"
>> ;-) I used to be one.
>> Ideally OMPI should do the switch between MXM/RC/XRC internally in the transport layer. Unfortunately,
>> we don't have such smart selection logic. Hopefully IB vendors will fix some day.
> I actually looked in the openib-hca.ini (working from memory) to try and find what the default queues were, and I actually couldn't figure it out. The ConnectX entry doesn't have a default, and the 'default default' also doesn't have an entry.
> I need to dig into ompi_info, got distracted by an intel compiler bug, ADD for admin/user support folks.
"default default" QP configuration is tuned for Mellanox devices.
Therefore the openib-hca.ini file doesn't have a special configuration for Connect-X.
The proper way to check default configuration is ompi_info utility.
>>> Note most of our users run just fine with the standard Peer-Peer queues, default out the box OpenMPI.
>> The P2P queue is fine, but most like using XRC your users will observe better performance. This is not just scalability.
> Cool thanks for all the input, I wonder why peer-to-peer is the default, I know XRC requires hardware support,
There is a historical reason behind this. OpenFabrics decided not to include XRC transport in default distribution. The XRC feature was only available as a part of
the Mellanox OFED distribution. I think, recently OFA community decided to include XRC, so we actually should consider to enable XRC by default.