Open MPI logo

Open MPI User's Mailing List Archives

  |   Home   |   Support   |   FAQ   |   all Open MPI User's mailing list

Subject: Re: [OMPI users] Socket ports
From: Jeff Squyres (jsquyres_at_[hidden])
Date: 2011-11-09 07:57:12


I could swear that we had an FAQ entry about this, but I can't find it.

It is certainly easiest if you can open random TCP ports between MPI processes in your cluster. Can your admin open all inbound TCP ports from all nodes in your cluster (this is different than opening up all inbound TCP ports from any IP address)?

Otherwise, you can try using 4 MCA params:

btl_tcp_port_min_v4
btl_tcp_port_range_v4
oob_tcp_port_min_v4
oob_tcp_port_range_v4

The first two control the ports used by MPI communications; the second two control Open MPI's "out of band" command/control messaging. The help messages for the two BTL params are:

mca:btl:tcp:param:btl_tcp_port_min_v4:value:1024
mca:btl:tcp:param:btl_tcp_port_min_v4:data_source:default value
mca:btl:tcp:param:btl_tcp_port_min_v4:status:writable
mca:btl:tcp:param:btl_tcp_port_min_v4:help:The minimum port where the TCP BTL will try to bind (default 1024)
mca:btl:tcp:param:btl_tcp_port_min_v4:deprecated:no
mca:btl:tcp:param:btl_tcp_port_range_v4:value:64511
mca:btl:tcp:param:btl_tcp_port_range_v4:data_source:default value
mca:btl:tcp:param:btl_tcp_port_range_v4:status:writable
mca:btl:tcp:param:btl_tcp_port_range_v4:help:The number of ports where the TCP BTL will try to bind (default 64511). This parameter together with the port min, define a range of ports where Open MPI will open sockets.
mca:btl:tcp:param:btl_tcp_port_range_v4:deprecated:no

Also, if you're running Open MPI v1.4.0, that's pretty old. You might want to upgrade; the latest stable version is 1.4.4.

On Nov 8, 2011, at 1:31 PM, Jeffrey A Cummings wrote:

> I'm attempting to launch my app via mpirun and a host file to use nodes on multiple 'stand-alone' servers. mpirun is able to launch my app on all requested nodes on all servers, but my app doesn't seem to be able to communicate via the standard MPI api calls (send , recv, etc). The problem seems to be that my sysadmin dept has locked down most/all ports for simple socket connections. They are asking me which specific ports (or range of ports) are required by mpi. I'm assuming that mpirun used secure sockets to launch my app on all nodes but that my app is not using secure sockets via the MPI calls. Does any of this make sense? I'm using version 1.4.0 I think.
>
> - Jeff Cummings_______________________________________________
> users mailing list
> users_at_[hidden]
> http://www.open-mpi.org/mailman/listinfo.cgi/users

-- 
Jeff Squyres
jsquyres_at_[hidden]
For corporate legal information go to:
http://www.cisco.com/web/about/doing_business/legal/cri/