Open MPI logo

Open MPI User's Mailing List Archives

  |   Home   |   Support   |   FAQ   |   all Open MPI User's mailing list

From: Justin Bronder (jsbronder_at_[hidden])
Date: 2006-07-06 15:15:35


With 1.0.3a1r10670 the same problem is occuring. Again the same configure
arguments
as before. For clarity, the Myrinet drive we are using is 2.0.21

node90:~/src/hpl/bin/ompi-xl-1.0.3 jbronder$ gm_board_info
GM build ID is "2.0.21_MacOSX_rc20050429075134PDT
root_at_[hidden]:/usr/src/gm-2.0.21_MacOSX Fri Jun 16
14:39:45 EDT 2006."

node90:~/src/hpl/bin/ompi-xl-1.0.3 jbronder$
/usr/local/ompi-xl-1.0.3/bin/mpirun
-np 2 xhpl
This succeeds.
||Ax-b||_oo / ( eps * ||A||_1 * N ) = 0.1196787 ...... PASSED
||Ax-b||_oo / ( eps * ||A||_1 * ||x||_1 ) = 0.0283195 ...... PASSED
||Ax-b||_oo / ( eps * ||A||_oo * ||x||_oo ) = 0.0063300 ...... PASSED

node90:~/src/hpl/bin/ompi-xl-1.0.3 jbronder$
/usr/local/ompi-xl-1.0.3/bin/mpirun
-mca btl gm -np 2 xhpl
This fails.
||Ax-b||_oo / ( eps * ||A||_1 * N ) =
717370209518881444284334080.0000000 ...... FAILED
||Ax-b||_oo / ( eps * ||A||_1 * ||x||_1 ) = 226686309135.4274597 ......
FAILED
||Ax-b||_oo / ( eps * ||A||_oo * ||x||_oo ) = 2386641249.6518722 ......
FAILED
||Ax-b||_oo . . . . . . . . . . . . . . . . . = 2037398812542965504.000000
||A||_oo . . . . . . . . . . . . . . . . . . . = 2561.554752
||A||_1 . . . . . . . . . . . . . . . . . . . = 2558.129237
||x||_oo . . . . . . . . . . . . . . . . . . . = 300175355203841216.000000
||x||_1 . . . . . . . . . . . . . . . . . . . = 31645943341479366656.000000

Does anyone have a working system with OS X and Myrinet (GM)? If so, I'd
love to hear
the configure arguments and various versions you are using. Bonus points if
you are
using the IBM XL compilers.

Thanks,
Justin.

On 7/6/06, Justin Bronder <jsbronder_at_[hidden]> wrote:
>
> Yes, that output was actually cut and pasted from an OS X run. I'm about
> to test
> against 1.0.3a1r10670.
>
> Justin.
>
> On 7/6/06, Galen M. Shipman <gshipman_at_[hidden]> wrote:
>
> > Justin,
> > Is the OS X run showing the same residual failure?
> >
> > - Galen
> >
> > On Jul 6, 2006, at 10:49 AM, Justin Bronder wrote:
> >
> > Disregard the failure on Linux, a rebuild from scratch of HPL and
> > OpenMPI
> > seems to have resolved the issue. At least I'm not getting the errors
> > during
> > the residual checks.
> >
> > However, this is persisting under OS X.
> >
> > Thanks,
> > Justin.
> >
> > On 7/6/06, Justin Bronder < jsbronder_at_[hidden]> wrote:
> >
> > > For OS X:
> > > /usr/local/ompi-xl/bin/mpirun -mca btl gm -np 4 ./xhpl
> > >
> > > For Linux:
> > > ARCH=ompi-gnu-1.1.1a
> > > /usr/local/$ARCH/bin/mpiexec -mca btl gm -np 2 -path
> > > /usr/local/$ARCH/bin ./xhpl
> > >
> > > Thanks for the speedy response,
> > > Justin.
> > >
> > > On 7/6/06, Galen M. Shipman < gshipman_at_[hidden]> wrote:
> > >
> > > > Hey Justin,
> > > Please provide us your mca parameters (if any), these could be in a
> > > config file, environment variables or on the command line.
> > >
> > > Thanks,
> > >
> > > Galen
> > >
> > > On Jul 6, 2006, at 9:22 AM, Justin Bronder wrote:
> > >
> > > As far as the nightly builds go, I'm still seeing what I believe to be
> > >
> > > this problem in both r10670 and r10652. This is happening with
> > > both Linux and OS X. Below are the systems and ompi_info for the
> > > newest revision 10670.
> > >
> > > As an example of the error, when running HPL with Myrinet I get the
> > > following error. Using tcp everything is fine and I see the results
> > > I'd
> > > expect.
> > >
> > > ----------------------------------------------------------------------------
> > > ||Ax-b||_oo / ( eps * ||A||_1 * N ) =
> > > 42820214496954887558164928727596662784.0000000 ...... FAILED
> > > ||Ax-b||_oo / ( eps * ||A||_1 * ||x||_1 ) = 156556068835.2711182...... FAILED
> > > ||Ax-b||_oo / ( eps * ||A||_oo * ||x||_oo ) = 1156439380.5172558...... FAILED
> > > ||Ax-b||_oo . . . . . . . . . . . . . . . . . =
> > > 272683853978565028754868928512.000000
> > > ||A||_oo . . . . . . . . . . . . . . . . . . . = 3822.884181
> > > ||A||_1 . . . . . . . . . . . . . . . . . . . = 3823.922627
> > > ||x||_oo . . . . . . . . . . . . . . . . . . . =
> > > 37037692483529688659798261760.000000
> > > ||x||_1 . . . . . . . . . . . . . . . . . . . =
> > > 4102704048669982798475494948864.000000
> > > ===================================================
> > >
> > > Finished 1 tests with the following results:
> > > 0 tests completed and passed residual checks,
> > > 1 tests completed and failed residual checks,
> > > 0 tests skipped because of illegal input values.
> > >
> > > ----------------------------------------------------------------------------
> > >
> > > Linux node41 2.6.16.19 #1 SMP Wed Jun 21 17:22:01 EDT 2006 ppc64
> > > PPC970FX, altivec supported GNU/Linux
> > > jbronder_at_node41 ~ $ /usr/local/ompi- gnu-1.1.1a/bin/ompi_info
> > > Open MPI: 1.1.1a1r10670
> > > Open MPI SVN revision: r10670
> > > Open RTE: 1.1.1a1r10670
> > > Open RTE SVN revision: r10670
> > > OPAL: 1.1.1a1r10670
> > > OPAL SVN revision: r10670
> > > Prefix: /usr/local/ompi-gnu-1.1.1a
> > > Configured architecture: powerpc64-unknown-linux-gnu
> > > Configured by: root
> > > Configured on: Thu Jul 6 10:15:37 EDT 2006
> > > Configure host: node41
> > > Built by: root
> > > Built on: Thu Jul 6 10:28:14 EDT 2006
> > > Built host: node41
> > > C bindings: yes
> > > C++ bindings: yes
> > > Fortran77 bindings: yes (all)
> > > Fortran90 bindings: yes
> > > Fortran90 bindings size: small
> > > C compiler: gcc
> > > C compiler absolute: /usr/bin/gcc
> > > C++ compiler: g++
> > > C++ compiler absolute: /usr/bin/g++
> > > Fortran77 compiler: gfortran
> > > Fortran77 compiler abs:
> > > /usr/powerpc64-unknown-linux-gnu/gcc-bin/4.1.0/gfortran
> > > Fortran90 compiler: gfortran
> > > Fortran90 compiler abs:
> > > /usr/powerpc64-unknown-linux-gnu/gcc-bin/4.1.0/gfortran
> > > C profiling: yes
> > > C++ profiling: yes
> > > Fortran77 profiling: yes
> > > Fortran90 profiling: yes
> > > C++ exceptions: no
> > > Thread support: posix (mpi: no, progress: no)
> > > Internal debug support: no
> > > MPI parameter check: runtime
> > > Memory profiling support: no
> > > Memory debugging support: no
> > > libltdl support: yes
> > > MCA memory: ptmalloc2 (MCA v1.0, API v1.0, Component
> > > v1.1.1)
> > > MCA paffinity: linux (MCA v1.0, API v1.0, Component v1.1.1)
> > > MCA maffinity: first_use (MCA v1.0, API v1.0, Component
> > > v1.1.1)
> > > MCA timer: linux (MCA v1.0, API v1.0, Component v1.1.1)
> > > MCA allocator: basic (MCA v1.0, API v1.0, Component v1.0)
> > > MCA allocator: bucket (MCA v1.0, API v1.0, Component v1.0)
> > > MCA coll: basic (MCA v1.0, API v1.0, Componentv1.1.1)
> > >
> > > MCA coll: hierarch (MCA v1.0, API v1.0, Component v1.1.1)
> > > MCA coll: self (MCA v1.0, API v1.0, Component v1.1.1)
> > > MCA coll: sm (MCA v1.0, API v1.0, Component v1.1.1)
> > > MCA coll: tuned (MCA v1.0, API v1.0, Component v1.1.1)
> > > MCA io: romio (MCA v1.0, API v1.0, Component v1.1.1)
> > > MCA mpool: gm (MCA v1.0, API v1.0, Component v1.1.1)
> > > MCA mpool: sm (MCA v1.0, API v1.0, Component v1.1.1)
> > > MCA pml: ob1 (MCA v1.0, API v1.0, Component v1.1.1)
> > > MCA bml: r2 (MCA v1.0, API v1.0, Component v1.1.1)
> > > MCA rcache: rb (MCA v1.0, API v1.0, Component v1.1.1)
> > > MCA btl: gm (MCA v1.0, API v1.0, Component v1.1.1)
> > > MCA btl: self (MCA v1.0, API v1.0, Component v1.1.1)
> > > MCA btl: sm (MCA v1.0, API v1.0, Component v1.1.1)
> > > MCA btl: tcp (MCA v1.0, API v1.0, Component v1.0)
> > > MCA topo: unity (MCA v1.0, API v1.0, Component v1.1.1)
> > > MCA osc: pt2pt (MCA v1.0, API v1.0, Component v1.0)
> > > MCA gpr: null (MCA v1.0, API v1.0, Component v1.1.1)
> > > MCA gpr: proxy (MCA v1.0, API v1.0, Component v1.1.1)
> > > MCA gpr: replica (MCA v1.0, API v1.0, Component
> > > v1.1.1)
> > > MCA iof: proxy (MCA v1.0, API v1.0, Component v1.1.1)
> > > MCA iof: svc (MCA v1.0, API v1.0, Component v1.1.1)
> > > MCA ns: proxy (MCA v1.0, API v1.0, Component v1.1.1)
> > > MCA ns: replica (MCA v1.0, API v1.0, Component
> > > v1.1.1)
> > > MCA oob: tcp (MCA v1.0, API v1.0, Component v1.0)
> > > MCA ras: dash_host (MCA v1.0, API v1.0, Component
> > > v1.1.1)
> > > MCA ras: hostfile (MCA v1.0, API v1.0, Component
> > > v1.1.1)
> > > MCA ras: localhost (MCA v1.0, API v1.0, Component
> > > v1.1.1)
> > > MCA ras: tm (MCA v1.0, API v1.0, Component v1.1.1)
> > > MCA rds: hostfile (MCA v1.0, API v1.0, Component
> > > v1.1.1)
> > > MCA rds: resfile (MCA v1.0, API v1.0, Component v1.1.1
> > > )
> > > MCA rmaps: round_robin (MCA v1.0, API v1.0, Component
> > > v1.1.1)
> > > MCA rmgr: proxy (MCA v1.0, API v1.0, Component v1.1.1)
> > > MCA rmgr: urm (MCA v1.0, API v1.0, Component v1.1.1)
> > > MCA rml: oob (MCA v1.0, API v1.0, Component v1.1.1)
> > > MCA pls: fork (MCA v1.0, API v1.0, Component v1.1.1)
> > > MCA pls: rsh (MCA v1.0, API v1.0, Component v1.1.1)
> > > MCA pls: tm (MCA v1.0, API v1.0, Component v1.1.1)
> > > MCA sds: env (MCA v1.0, API v1.0, Component v1.1.1)
> > > MCA sds: pipe (MCA v1.0, API v1.0, Component v1.1.1)
> > > MCA sds: seed (MCA v1.0, API v1.0, Component v1.1.1)
> > > MCA sds: singleton (MCA v1.0, API v1.0, Component
> > > v1.1.1)
> > > Configured as:
> > > ./configure \
> > > --prefix=$PREFIX \
> > > --enable-mpi-f77 \
> > > --enable-mpi-f90 \
> > > --enable-mpi-profile \
> > > --enable-mpi-cxx \
> > > --enable-pty-support \
> > > --enable-shared \
> > > --enable-smp-locks \
> > > --enable-io-romio \
> > > --with-tm=/usr/local/pbs \
> > > --without-xgrid \
> > > --without-slurm \
> > > --with-gm=/opt/gm
> > >
> > > Darwin node90.meldrew.clusters.umaine.edu 8.6.0 Darwin Kernel Version
> > > 8.6.0: Tue Mar 7 16:58:48 PST 2006; root:
> > > xnu-792.6.70.obj~1/RELEASE_PPC Power Macintosh powerpc
> > > node90:~/src/hpl jbronder$ /usr/local/ompi-xl/bin/ompi_info
> > > Open MPI: 1.1.1a1r10670
> > > Open MPI SVN revision: r10670
> > > Open RTE: 1.1.1a1r10670
> > > Open RTE SVN revision: r10670
> > > OPAL: 1.1.1a1r10670
> > > OPAL SVN revision: r10670
> > > Prefix: /usr/local/ompi-xl
> > > Configured architecture: powerpc-apple-darwin8.6.0
> > > Configured by:
> > > Configured on: Thu Jul 6 10:05:20 EDT 2006
> > > Configure host: node90.meldrew.clusters.umaine.edu
> > > Built by: root
> > > Built on: Thu Jul 6 10:37:40 EDT 2006
> > > Built host: node90.meldrew.clusters.umaine.edu
> > > C bindings: yes
> > > C++ bindings: yes
> > > Fortran77 bindings: yes (lower case)
> > > Fortran90 bindings: yes
> > > Fortran90 bindings size: small
> > > C compiler: /opt/ibmcmp/vac/6.0/bin/xlc
> > > C compiler absolute: /opt/ibmcmp/vac/6.0/bin/xlc
> > > C++ compiler: /opt/ibmcmp/vacpp/6.0/bin/xlc++
> > > C++ compiler absolute: /opt/ibmcmp/vacpp/6.0/bin/xlc++
> > > Fortran77 compiler: /opt/ibmcmp/xlf/8.1/bin/xlf_r
> > > Fortran77 compiler abs: /opt/ibmcmp/xlf/8.1/bin/xlf_r
> > > Fortran90 compiler: /opt/ibmcmp/xlf/8.1/bin/xlf90_r
> > > Fortran90 compiler abs: /opt/ibmcmp/xlf/8.1/bin/xlf90_r
> > > C profiling: yes
> > > C++ profiling: yes
> > > Fortran77 profiling: yes
> > > Fortran90 profiling: yes
> > > C++ exceptions: no
> > > Thread support: posix (mpi: no, progress: no)
> > > Internal debug support: no
> > > MPI parameter check: runtime
> > > Memory profiling support: no
> > > Memory debugging support: no
> > > libltdl support: yes
> > > MCA memory: darwin (MCA v1.0, API v1.0, Component v1.1.1
> > > )
> > > MCA maffinity: first_use (MCA v1.0, API v1.0, Component
> > > v1.1.1)
> > > MCA timer: darwin (MCA v1.0, API v1.0, Component v1.1.1
> > > )
> > > MCA allocator: basic (MCA v1.0, API v1.0, Component v1.0)
> > > MCA allocator: bucket (MCA v1.0, API v1.0, Component v1.0)
> > > MCA coll: basic (MCA v1.0, API v1.0, Component v1.1.1)
> > > MCA coll: hierarch (MCA v1.0, API v1.0, Component
> > > v1.1.1)
> > > MCA coll: self (MCA v1.0, API v1.0, Component v1.1.1)
> > > MCA coll: sm (MCA v1.0, API v1.0, Component v1.1.1)
> > > MCA coll: tuned (MCA v1.0, API v1.0, Component v1.1.1)
> > > MCA io: romio (MCA v1.0, API v1.0, Component v1.1.1)
> > > MCA mpool: sm (MCA v1.0, API v1.0, Component v1.1.1)
> > > MCA mpool: gm (MCA v1.0, API v1.0, Component v1.1.1)
> > > MCA pml: ob1 (MCA v1.0, API v1.0, Component v1.1.1)
> > > MCA bml: r2 (MCA v1.0, API v1.0, Component v1.1.1)
> > > MCA rcache: rb (MCA v1.0, API v1.0, Component v1.1.1)
> > > MCA btl: self (MCA v1.0, API v1.0, Component v1.1.1)
> > > MCA btl: sm (MCA v1.0, API v1.0, Component v1.1.1)
> > > MCA btl: gm (MCA v1.0, API v1.0, Component v1.1.1)
> > > MCA btl: tcp (MCA v1.0, API v1.0, Component v1.0)
> > > MCA topo: unity (MCA v1.0, API v1.0, Component v1.1.1)
> > > MCA osc: pt2pt (MCA v1.0, API v1.0, Component v1.0)
> > > MCA gpr: null (MCA v1.0, API v1.0, Component v1.1.1)
> > > MCA gpr: proxy (MCA v1.0, API v1.0, Component v1.1.1)
> > > MCA gpr: replica (MCA v1.0, API v1.0, Component
> > > v1.1.1)
> > > MCA iof: proxy (MCA v1.0, API v1.0, Component v1.1.1)
> > > MCA iof: svc (MCA v1.0, API v1.0, Component v1.1.1)
> > > MCA ns: proxy (MCA v1.0, API v1.0, Component v1.1.1)
> > > MCA ns: replica (MCA v1.0, API v1.0, Component
> > > v1.1.1)
> > > MCA oob: tcp (MCA v1.0, API v1.0, Component v1.0)
> > > MCA ras: dash_host (MCA v1.0, API v1.0, Component
> > > v1.1.1)
> > > MCA ras: hostfile (MCA v1.0, API v1.0, Component
> > > v1.1.1)
> > > MCA ras: localhost (MCA v1.0, API v1.0, Component
> > > v1.1.1)
> > > MCA ras: tm (MCA v1.0, API v1.0, Component v1.1.1)
> > > MCA rds: hostfile (MCA v1.0, API v1.0, Component v1.1.1
> > > )
> > > MCA rds: resfile (MCA v1.0, API v1.0, Component
> > > v1.1.1)
> > > MCA rmaps: round_robin (MCA v1.0, API v1.0, Component
> > > v1.1.1)
> > > MCA rmgr: proxy (MCA v1.0, API v1.0, Component v1.1.1)
> > > MCA rmgr: urm (MCA v1.0, API v1.0, Component v1.1.1)
> > > MCA rml: oob (MCA v1.0, API v1.0, Component v1.1.1)
> > > MCA pls: fork (MCA v1.0, API v1.0, Component v1.1.1)
> > > MCA pls: rsh (MCA v1.0, API v1.0, Component v1.1.1)
> > > MCA pls: tm (MCA v1.0, API v1.0, Component v1.1.1)
> > > MCA sds: env (MCA v1.0, API v1.0, Component v1.1.1)
> > > MCA sds: seed (MCA v1.0, API v1.0, Component v1.1.1)
> > > MCA sds: singleton (MCA v1.0, API v1.0, Component
> > > v1.1.1)
> > > MCA sds: pipe (MCA v1.0, API v1.0, Component v1.1.1)
> > > Configured as:
> > > ./configure \
> > > --prefix=$PREFIX \
> > > --with-tm=/usr/local/pbs/ \
> > > --with-gm=/opt/gm \
> > > --enable-static \
> > > --disable-cxx
> > > On 7/3/06, George Bosilca < bosilca_at_[hidden] > wrote:
> > >
> > > > Bernard,
> > >
> > > A bug in the Open MPI GM driver was discovered after the 1.1 release.
> > > A patch for the 1.1 is on the way. However, I don't know if it will
> > > be available before the 1.1.1. Meanwhile, you can use the nightly
> > > build version or a fresh check-out from the SVN repository. Both of
> > > them have the GM bug corrected.
> > >
> > > Sorry for the troubles,
> > > george.
> > >
> > > On Jul 3, 2006, at 12:58 PM, Borenstein, Bernard S wrote:
> > >
> > > > I've built and sucessfully run the Nasa Overflow 2.0aa program with
> > > > Openmpi 1.0.2. I'm running on an opteron linux cluster running SLES
> > > 9
> > > > and GM 2.0.24. I built Openmpi 1.1 with the intel 9 compilers and
> > > > try to
> > > > run Overflow 2.0aa with myrinet, it get what looks like a data
> > > > corruption error and the program dies quickly.
> > > > There are no mpi errors at all.If I run using GIGE (--mca btl
> > > > self,tcp),
> > > > the program runs to competion correctly. Here is my ompi_info
> > > > output :
> > > >
> > > > bsb3227_at_mahler:~/openmpi_1.1/bin> ./ompi_info
> > > > Open MPI: 1.1
> > > > Open MPI SVN revision: r10477
> > > > Open RTE: 1.1
> > > > Open RTE SVN revision: r10477
> > > > OPAL: 1.1
> > > > OPAL SVN revision: r10477
> > > > Prefix: /home/bsb3227/openmpi_1.1
> > > > Configured architecture: x86_64-unknown-linux-gnu
> > > > Configured by: bsb3227
> > > > Configured on: Fri Jun 30 07:08:54 PDT 2006
> > > > Configure host: mahler
> > > > Built by: bsb3227
> > > > Built on: Fri Jun 30 07:54:46 PDT 2006
> > > > Built host: mahler
> > > > C bindings: yes
> > > > C++ bindings: yes
> > > > Fortran77 bindings: yes (all)
> > > > Fortran90 bindings: yes
> > > > Fortran90 bindings size: small
> > > > C compiler: icc
> > > > C compiler absolute: /opt/intel/cce/9.0.25/bin/icc
> > > > C++ compiler: icpc
> > > > C++ compiler absolute: /opt/intel/cce/9.0.25/bin/icpc
> > > > Fortran77 compiler: ifort
> > > > Fortran77 compiler abs: /opt/intel/fce/9.0.25/bin/ifort
> > > > Fortran90 compiler: /opt/intel/fce/9.0.25/bin/ifort
> > > > Fortran90 compiler abs: /opt/intel/fce/9.0.25/bin/ifort
> > > > C profiling: yes
> > > > C++ profiling: yes
> > > > Fortran77 profiling: yes
> > > > Fortran90 profiling: yes
> > > > C++ exceptions: no
> > > > Thread support: posix (mpi: no, progress: no)
> > > > Internal debug support: no
> > > > MPI parameter check: runtime
> > > > Memory profiling support: no
> > > > Memory debugging support: no
> > > > libltdl support: yes
> > > > MCA memory: ptmalloc2 (MCA v1.0, API v1.0, Component
> > > > v1.1)
> > > > MCA paffinity: linux (MCA v1.0, API v1.0, Component v1.1)
> > > > MCA maffinity: first_use (MCA v1.0, API v1.0, Component
> > > > v1.1)
> > > > MCA maffinity: libnuma (MCA v1.0, API v1.0, Component
> > > v1.1)
> > > > MCA timer: linux (MCA v1.0, API v1.0, Component v1.1)
> > > > MCA allocator: basic (MCA v1.0, API v1.0, Component v1.0)
> > > > MCA allocator: bucket (MCA v1.0, API v1.0, Component v1.0
> > > )
> > > > MCA coll: basic (MCA v1.0, API v1.0, Component v1.1)
> > > > MCA coll: hierarch (MCA v1.0, API v1.0, Component
> > > > v1.1)
> > > > MCA coll: self (MCA v1.0, API v1.0, Component v1.1)
> > > > MCA coll: sm (MCA v1.0, API v1.0, Component v1.1)
> > > > MCA coll: tuned (MCA v1.0, API v1.0, Component v1.1)
> > > > MCA io: romio (MCA v1.0, API v1.0, Component v1.1)
> > > > MCA mpool: sm (MCA v1.0, API v1.0, Component v1.1)
> > > > MCA mpool: gm (MCA v1.0, API v1.0, Component v1.1)
> > > > MCA pml: ob1 (MCA v1.0, API v1.0, Component v1.1)
> > > > MCA bml: r2 (MCA v1.0, API v1.0, Component v1.1)
> > > > MCA rcache: rb (MCA v1.0, API v1.0, Component v1.1)
> > > > MCA btl: self (MCA v1.0, API v1.0, Component v1.1)
> > > > MCA btl: sm (MCA v1.0, API v1.0, Component v1.1)
> > > > MCA btl: gm (MCA v1.0, API v1.0, Component v1.1)
> > > > MCA btl: tcp (MCA v1.0, API v1.0, Component v1.0)
> > > > MCA topo: unity (MCA v1.0, API v1.0, Component v1.1)
> > > > MCA osc: pt2pt (MCA v1.0, API v1.0, Component v1.0)
> > > > MCA gpr: null (MCA v1.0, API v1.0, Component v1.1)
> > > > MCA gpr: proxy (MCA v1.0, API v1.0, Component v1.1)
> > > > MCA gpr: replica (MCA v1.0, API v1.0, Component
> > > v1.1)
> > > > MCA iof: proxy (MCA v1.0, API v1.0, Component v1.1)
> > > > MCA iof: svc (MCA v1.0, API v1.0, Component v1.1)
> > > > MCA ns: proxy (MCA v1.0, API v1.0, Component v1.1)
> > > > MCA ns: replica (MCA v1.0, API v1.0, Component
> > > v1.1)
> > > > MCA oob: tcp (MCA v1.0, API v1.0, Component v1.0)
> > > > MCA ras: dash_host (MCA v1.0, API v1.0, Component
> > > > v1.1)
> > > > MCA ras: hostfile (MCA v1.0, API v1.0, Component
> > > > v1.1)
> > > > MCA ras: localhost (MCA v1.0, API v1.0, Component
> > > > v1.1)
> > > > MCA ras: slurm (MCA v1.0, API v1.0, Component v1.1)
> > > > MCA ras: tm (MCA v1.0, API v1.0, Component v1.1)
> > > > MCA rds: hostfile (MCA v1.0, API v1.0, Component
> > > > v1.1)
> > > > MCA rds: resfile (MCA v1.0, API v1.0, Component
> > > v1.1)
> > > > MCA rmaps: round_robin (MCA v1.0, API v1.0, Component
> > > > v1.1)
> > > > MCA rmgr: proxy (MCA v1.0, API v1.0, Component v1.1)
> > > > MCA rmgr: urm (MCA v1.0, API v1.0, Component v1.1)
> > > > MCA rml: oob (MCA v1.0, API v1.0, Component v1.1)
> > > > MCA pls: fork (MCA v1.0, API v1.0, Component v1.1)
> > > > MCA pls: rsh (MCA v1.0, API v1.0, Component v1.1)
> > > > MCA pls: slurm (MCA v1.0, API v1.0, Component v1.1)
> > > > MCA pls: tm (MCA v1.0, API v1.0, Component v1.1)
> > > > MCA sds: env (MCA v1.0, API v1.0, Component v1.1)
> > > > MCA sds: seed (MCA v1.0, APIv1.0, Component v1.1)
> > >
> > > > MCA sds: singleton (MCA v1.0, API v1.0, Component
> > > > v1.1)
> > > > MCA sds: pipe (MCAv1.0, API v1.0, Component v1.1)
> > >
> > > > MCA sds: slurm (MCA v1.0, API v1.0, Component v1.1)
> > > >
> > > > Here is the ifconfig for one of the nodes :
> > > >
> > > > bsb3227_at_m045:~> /sbin/ifconfig
> > > > eth0 Link encap:Ethernet HWaddr 00:50:45:5D:CD:FE
> > > > inet addr:10.241.194.45 Bcast: 10.241.195.255
> > > > Mask:255.255.254.0
> > > > inet6 addr: fe80::250:45ff:fe5d:cdfe/64 Scope:Link
> > > > UP BROADCAST NOTRAILERS RUNNING MULTICAST MTU:1500
> > > > Metric:1
> > > > RX packets:39913407 errors:0 dropped:0 overruns:0 frame:0
> > > > TX packets:48794587 errors:0 dropped:0 overruns:0
> > > carrier:0
> > > > collisions:0 txqueuelen:1000
> > > > RX bytes:31847343907 (30371.9 Mb) TX bytes:48231713866
> > > > (45997.3 Mb)
> > > > Interrupt:19
> > > >
> > > > eth1 Link encap:Ethernet HWaddr 00:50:45:5D:CD:FF
> > > > inet6 addr: fe80::250:45ff:fe5d:cdff/64 Scope:Link
> > > > UP BROADCAST MULTICAST MTU:1500 Metric:1
> > > > RX packets:0 errors:0 dropped:0 overruns:0 frame:0
> > > > TX packets:0 errors:0 dropped:0 overruns:0 carrier:0
> > > > collisions:0 txqueuelen:1000
> > > > RX bytes:0 (0.0 b) TX bytes:0 (0.0 b)
> > > > Interrupt:19
> > > >
> > > > lo Link encap:Local Loopback
> > > > inet addr: 127.0.0.1 Mask: 255.0.0.0
> > > > inet6 addr: ::1/128 Scope:Host
> > > > UP LOOPBACK RUNNING MTU:16436 Metric:1
> > > > RX packets:23141 errors:0 dropped:0 overruns:0 frame:0
> > > > TX packets:23141 errors:0 dropped:0 overruns:0 carrier:0
> > > > collisions:0 txqueuelen:0
> > > > RX bytes:20145689 (19.2 Mb) TX bytes:20145689 (19.2 Mb)
> > > >
> > > > I hope someone can give me some guidance on how to debug this
> > > problem.
> > > > Thanx in advance for any help
> > > > that can be provided.
> > > >
> > > > Bernie Borenstein
> > > > The Boeing Company
> > > > <config.log.gz>
> > > > _______________________________________________
> > > > users mailing list
> > > > users_at_[hidden]
> > > > http://www.open-mpi.org/mailman/listinfo.cgi/users
> > >
> > > "Half of what I say is meaningless; but I say it so that the other
> > > half may reach you"
> > > Kahlil Gibran
> > >
> > >
> > > _______________________________________________
> > > users mailing list
> > > users_at_[hidden]
> > > http://www.open-mpi.org/mailman/listinfo.cgi/users
> > >
> >
> > _______________________________________________
> > users mailing list
> > users_at_[hidden]
> > http://www.open-mpi.org/mailman/listinfo.cgi/users
> >
> >
> _______________________________________________
> users mailing list
> users_at_[hidden]
> http://www.open-mpi.org/mailman/listinfo.cgi/users
>
>