On Oct 16, 2009, at 1:55 PM, nam kim wrote:
> Our school has a cluster running over CISCO based Infiniband cards
> and switch.
> Recently, we purchased more computing nods with Mellanox card since
> CISCO stops making IB card anymore.
Sorry for the delay in replying; my INBOX has grown totally out of
hand recently. :-(
FWIW, Cisco never made IB HCAs; we simply resold Mellanox HCAs.
> Currently, I use openmpi 1.2.8 compiled with CISCO IB card (SFS-
> HCA-320-A1) with topspin driver. My questions are:
> 1. Is it possible to compile 1.3 version with mixed cisco IB and
> mellanox IB (MHRH19-XTC) with open infiniband libries?
Do you mean: is it possible to use Open MPI 1.3.x with a recent OFED
distribution across multiple nodes, some of which include Cisco-
branded HCAs and some of which include Mellanox HCAs?
The answer is: most likely, yes. Open MPI doesn't fully support
"heterogeneous" HCAs (e.g., HCAs that would require different MTUs).
But I suspect that your HCAs are all "close enough" that it won't
matter. FWIW, on my 64-node MPI testing cluster at Cisco, I do
similar things -- I have various Cisco and Mellanox HCAs of different
generations and specific capabilities, and Open MPI runs fine.
> 2. Is is possible to compile 1.2.8 with mixed cisco IB and mellanox
> IB, then how?
If you can, I'd highly suggest upgrading to the Open MPI v1.3 series.