Open MPI logo

Open MPI User's Mailing List Archives

  |   Home   |   Support   |   FAQ   |  

This web mail archive is frozen.

This page is part of a frozen web archive of this mailing list.

You can still navigate around this archive, but know that no new mails have been added to it since July of 2016.

Click here to be taken to the new web archives of this list; it includes all the mails that are in this frozen archive plus all new mails that have been sent to the list since it was migrated to the new archives.

From: Francoise Roch (roch_at_[hidden])
Date: 2006-05-29 07:46:57


hello,

I still have a problem to select an interface with openmpi-1.1a7 on
solaris opteron.
I compile in 64 bit mode, with Studio11 compilers

I attempted to force interface exclusion without success.
This problem is critical for us because we'll soon have Infiniband
interfaces for mpi traffic.

roch_at_n15 ~/MPI > mpirun --mca btl_tcp_if_exclude bge1 -np 2 -host
p15,p27 all2all
Process 0 is alive on n15
Process 1 is alive on n27
[n27:05110] *** An error occurred in MPI_Barrier
[n27:05110] *** on communicator MPI_COMM_WORLD
[n27:05110] *** MPI_ERR_INTERN: internal error
[n27:05110] *** MPI_ERRORS_ARE_FATAL (goodbye)
1 process killed (possibly by Open MPI)

The code works without mca btl_tcp_if_exclude option.

Currently each node has two ethernet interfaces :
roch_at_n15 ~ > /usr/sbin/ifconfig -a
lo0: flags=2001000849<UP,LOOPBACK,RUNNING,MULTICAST,IPv4,VIRTUAL> mtu
8232 index 1
         inet 127.0.0.1 netmask ff000000
bge0: flags=1000843<UP,BROADCAST,RUNNING,MULTICAST,IPv4> mtu 1500 index 2
         inet 192.168.12.15 netmask ffffff00 broadcast 192.168.12.255
bge1: flags=1004843<UP,BROADCAST,RUNNING,MULTICAST,DHCP,IPv4> mtu 1500
index 3
         inet 192.168.11.15 netmask ffffff00 broadcast 192.168.11.255

The hostname of the nodes are nxx.
in /etc/hosts, nxx corresponds to the interface bge1; pxx to the
interface bge0.

Thanks in advance for any help.

Françoise