Open MPI logo

Open MPI User's Mailing List Archives

  |   Home   |   Support   |   FAQ   |  

This web mail archive is frozen.

This page is part of a frozen web archive of this mailing list.

You can still navigate around this archive, but know that no new mails have been added to it since July of 2016.

Click here to be taken to the new web archives of this list; it includes all the mails that are in this frozen archive plus all new mails that have been sent to the list since it was migrated to the new archives.

From: Justin Bronder (jsbronder_at_[hidden])
Date: 2006-06-29 17:18:41

I'm having trouble getting OpenMPI to execute jobs when submitting through
Everything works fine if I am to use the included mpirun scripts, but this
is obviously
not a good solution for the general users on the cluster.

I'm running under OS X 10.4, Darwin 8.6.0. I configured OpenMpi with:
export CC=/opt/ibmcmp/vac/6.0/bin/xlc
export CXX=/opt/ibmcmp/vacpp/6.0/bin/xlc++
export FC=/opt/ibmcmp/xlf/8.1/bin/xlf90_r
export F77=/opt/ibmcmp/xlf/8.1/bin/xlf_r
export LDFLAGS=-lSystemStubs
export LIBTOOL=glibtool


./configure \
    --prefix=$PREFIX \
    --with-tm=/usr/local/pbs/ \
    --with-gm=/opt/gm \
    --enable-static \

I also had to employ the fix listed in:

I've attached the output of ompi_info while in an interactive job. Looking
through the list,
I can at least save a bit of trouble by listing what does work. Anything
outside of Torque
seems fine. From within an interactive job, pbsdsh works fine, hence the
earlier problems
with poll are fixed.

Here is the error that is reported when I attemt to run hostname on one
node96:/usr/src/openmpi-1.1 jbronder$ /usr/local/ompi-xl/bin/mpirun -np 1
-mca pls_tm_debug 1 /bin/hostname
[] pls:tm: final top-level argv:
[] pls:tm: orted --no-daemonize
--bootproxy 1 --name --num_procs 2 --vpid_start 0 --nodename --universe
jbronder_at_[hidden]:default-universe --nsreplica "
0.0.0;tcp://" --gprreplica "0.0.0;tcp://"
[] pls:tm: Set
[] pls:tm: launching on node
[] pls:tm: resetting PATH:
[] pls:tm: found
[] pls:tm: not oversubscribed --
setting mpi_yield_when_idle to 0
[] pls:tm: executing: orted
--no-daemonize --bootproxy 1 --name 0.0.1 --num_procs 2 --vpid_start 0
--nodename localhost --universe
--nsreplica "0.0.0;tcp://" --gprreplica "0.0.0
[] pls:tm: start_procs returned
error -13
[] [0,0,0] ORTE_ERROR_LOG: Not found
in file rmgr_urm.c at line 184
[] [0,0,0] ORTE_ERROR_LOG: Not found
in file rmgr_urm.c at line 432
[] mpirun: spawn failed with
node96:/usr/src/openmpi-1.1 jbronder$

My thanks for any help in advance,

Justin Bronder.