Open MPI logo

Open MPI User's Mailing List Archives

  |   Home   |   Support   |   FAQ   |   all Open MPI User's mailing list

From: Galen M. Shipman (gshipman_at_[hidden])
Date: 2006-07-11 16:15:00


Hey Justin,

Any further details?

Thanks,

Galen

On Jul 8, 2006, at 9:10 AM, Justin Bronder wrote:

> 1.) Compiling without XL will take a little while, but I have the
> setup
> for the
> other questions ready now. I figured I'd answer them right away.
>
> 2.) TCP works fine, and is quite quick compared to mpich-1.2.7p1
> by the
> way.
> I just reverified this.
> WR11C2R4 5000 160 1 2 10.10
> 8.253e+00
> ||Ax-b||_oo / ( eps * ||A||_1 * N ) =
> 0.0412956 ...... PASSED
> ||Ax-b||_oo / ( eps * ||A||_1 * ||x||_1 ) =
> 0.0272613 ...... PASSED
> ||Ax-b||_oo / ( eps * ||A||_oo * ||x||_oo ) =
> 0.0053214 ...... PASSED
>
>
> 3.) Exactly same setup, using mpichgm-1.2.6..14b
> WR11C2R4 5000 160 1 2 10.43
> 7.994e+00
> ----------------------------------------------------------------------
> ------
> ||Ax-b||_oo / ( eps * ||A||_1 * N ) =
> 0.0353693 ...... PASSED
> ||Ax-b||_oo / ( eps * ||A||_1 * ||x||_1 ) =
> 0.0233491 ...... PASSED
> ||Ax-b||_oo / ( eps * ||A||_oo * ||x||_oo ) =
> 0.0045577 ...... PASSED
>
> It also worked with mpichgm-1.2.6..15 (I believe this is the
> version, I
> don't have
> a node up with it at the moment).
>
> Obviously mpich-1.2.7p1 works as well over ethernet.
>
>
> Anyways, I'll begin the build with the standard gcc compilers that are
> included
> with OS X. This is powerpc-apple-darwin8-gcc-4.0.1.
>
> Thanks,
>
> Justin.
>
> Jeff Squyres (jsquyres) wrote:
>> Justin --
>>
>> Can we eliminate some variables so that we can figure out where the
>> error is originating?
>>
>> - Can you try compiling without the XL compilers?
>> - Can you try running with just TCP (and not Myrinet)?
>> - With the same support library installation (such as BLAS, etc.,
>> assumedly also compiled with XL), can you try another MPI (e.g., LAM,
>> MPICH-gm, whatever)?
>>
>> Let us know what you find. Thanks!
>>
>>
>>
>> ---------------------------------------------------------------------
>> ---
>> *From:* users-bounces_at_[hidden]
>> [mailto:users-bounces_at_[hidden]] *On Behalf Of *Justin Bronder
>> *Sent:* Thursday, July 06, 2006 3:16 PM
>> *To:* Open MPI Users
>> *Subject:* Re: [OMPI users] Problem with Openmpi 1.1
>>
>> With 1.0.3a1r10670 the same problem is occuring. Again the same
>> configure arguments
>> as before. For clarity, the Myrinet drive we are using is 2.0.21
>>
>> node90:~/src/hpl/bin/ompi-xl-1.0.3 jbronder$ gm_board_info
>> GM build ID is "2.0.21_MacOSX_rc20050429075134PDT
>> root_at_[hidden]:/usr/src/gm-2.0.21_MacOSX
>> Fri Jun 16 14:39:45 EDT 2006."
>>
>> node90:~/src/hpl/bin/ompi-xl-1.0.3 jbronder$
>> /usr/local/ompi-xl-1.0.3/bin/mpirun -np 2 xhpl
>> This succeeds.
>> ||Ax-b||_oo / ( eps * ||A||_1 * N ) = 0.1196787
>> ...... PASSED
>> ||Ax-b||_oo / ( eps * ||A||_1 * ||x||_1 ) = 0.0283195
>> ...... PASSED
>> ||Ax-b||_oo / ( eps * ||A||_oo * ||x||_oo ) = 0.0063300
>> ...... PASSED
>>
>> node90:~/src/hpl/bin/ompi-xl-1.0.3 jbronder$
>> /usr/local/ompi-xl-1.0.3/bin/mpirun -mca btl gm -np 2 xhpl
>> This fails.
>> ||Ax-b||_oo / ( eps * ||A||_1 * N ) =
>> 717370209518881444284334080.0000000 ...... FAILED
>> ||Ax-b||_oo / ( eps * ||A||_1 * ||x||_1 ) =
>> 226686309135.4274597
>> ...... FAILED
>> ||Ax-b||_oo / ( eps * ||A||_oo * ||x||_oo ) = 2386641249.6518722
>> ...... FAILED
>> ||Ax-b||_oo . . . . . . . . . . . . . . . . . =
>> 2037398812542965504.000000
>> ||A||_oo . . . . . . . . . . . . . . . . . . . =
>> 2561.554752
>> ||A||_1 . . . . . . . . . . . . . . . . . . . =
>> 2558.129237
>> ||x||_oo . . . . . . . . . . . . . . . . . . . =
>> 300175355203841216.000000
>> ||x||_1 . . . . . . . . . . . . . . . . . . . =
>> 31645943341479366656.000000
>>
>> Does anyone have a working system with OS X and Myrinet (GM)? If
>> so, I'd love to hear
>> the configure arguments and various versions you are using.
>> Bonus
>> points if you are
>> using the IBM XL compilers.
>>
>> Thanks,
>> Justin.
>>
>>
>> On 7/6/06, *Justin Bronder* <jsbronder_at_[hidden]
>> <mailto:jsbronder_at_[hidden]>> wrote:
>>
>> Yes, that output was actually cut and pasted from an OS X
>> run. I'm about to test
>> against 1.0.3a1r10670.
>>
>> Justin.
>>
>> On 7/6/06, *Galen M. Shipman* < gshipman_at_[hidden]
>> <mailto:gshipman_at_[hidden]>> wrote:
>>
>> Justin,
>>
>> Is the OS X run showing the same residual failure?
>>
>> - Galen
>>
>> On Jul 6, 2006, at 10:49 AM, Justin Bronder wrote:
>>
>> Disregard the failure on Linux, a rebuild from scratch of
>> HPL and OpenMPI
>> seems to have resolved the issue. At least I'm not
>> getting the errors during
>> the residual checks.
>>
>> However, this is persisting under OS X.
>>
>> Thanks,
>> Justin.
>>
>> On 7/6/06, *Justin Bronder* < jsbronder_at_[hidden]
>> <mailto:jsbronder_at_[hidden]>> wrote:
>>
>> For OS X:
>> /usr/local/ompi-xl/bin/mpirun -mca btl gm -np 4 ./
>> xhpl
>>
>> For Linux:
>> ARCH=ompi-gnu-1.1.1a
>> /usr/local/$ARCH/bin/mpiexec -mca btl gm -np 2 -path
>> /usr/local/$ARCH/bin ./xhpl
>>
>> Thanks for the speedy response,
>> Justin.
>>
>> On 7/6/06, *Galen M. Shipman* < gshipman_at_[hidden]
>> <mailto:gshipman_at_[hidden]>> wrote:
>> Hey Justin,
>>
>> Please provide us your mca parameters (if any), these
>> could be in a config file, environment variables
>> or on
>> the command line.
>>
>> Thanks,
>>
>> Galen
>>
>> On Jul 6, 2006, at 9:22 AM, Justin Bronder wrote:
>>
>> As far as the nightly builds go, I'm still seeing
>> what
>> I believe to be
>> this problem in both r10670 and r10652. This is
>> happening with
>> both Linux and OS X. Below are the systems and
>> ompi_info for the
>> newest revision 10670.
>>
>> As an example of the error, when running HPL with
>> Myrinet I get the
>> following error. Using tcp everything is fine and I
>> see the results I'd
>> expect.
>>
>> ---------------------------------------------------------------------
>> -------
>> ||Ax-b||_oo / ( eps * ||A||_1 * N ) =
>> 42820214496954887558164928727596662784.0000000 ......
>> FAILED
>> ||Ax-b||_oo / ( eps * ||A||_1 * ||x||_1 ) =
>> 156556068835.2711182 ...... FAILED
>> ||Ax-b||_oo / ( eps * ||A||_oo * ||x||_oo ) =
>> 1156439380.5172558 ...... FAILED
>> ||Ax-b||_oo . . . . . . . . . . . . . . . . . =
>> 272683853978565028754868928512.000000
>> ||A||_oo . . . . . . . . . . . . . . . . . . .
>> = 3822.884181
>> ||A||_1 . . . . . . . . . . . . . . . . . . .
>> = 3823.922627
>> ||x||_oo . . . . . . . . . . . . . . . . . . . =
>> 37037692483529688659798261760.000000
>> ||x||_1 . . . . . . . . . . . . . . . . . . . =
>> 4102704048669982798475494948864.000000
>> ===================================================
>>
>> Finished 1 tests with the following results:
>> 0 tests completed and passed residual
>> checks,
>> 1 tests completed and failed residual
>> checks,
>> 0 tests skipped because of illegal
>> input
>> values.
>>
>> ---------------------------------------------------------------------
>> -------
>>
>> Linux node41 2.6.16.19 <http://2.6.16.19> #1 SMP Wed
>> Jun 21 17:22:01 EDT 2006 ppc64 PPC970FX, altivec
>> supported GNU/Linux
>> jbronder_at_node41 ~ $ /usr/local/ompi-
>> gnu-1.1.1a/bin/ompi_info
>> Open MPI: 1.1.1a1r10670
>> Open MPI SVN revision: r10670
>> Open RTE: 1.1.1a1r10670
>> Open RTE SVN revision: r10670
>> OPAL: 1.1.1a1r10670
>> OPAL SVN revision: r10670
>> Prefix: /usr/local/ompi-gnu-1.1.1a
>> Configured architecture: powerpc64-unknown-linux-gnu
>> Configured by: root
>> Configured on: Thu Jul 6 10:15:37 EDT
>> 2006
>> Configure host: node41
>> Built by: root
>> Built on: Thu Jul 6 10:28:14 EDT
>> 2006
>> Built host: node41
>> C bindings: yes
>> C++ bindings: yes
>> Fortran77 bindings: yes (all)
>> Fortran90 bindings: yes
>> Fortran90 bindings size: small
>> C compiler: gcc
>> C compiler absolute: /usr/bin/gcc
>> C++ compiler: g++
>> C++ compiler absolute: /usr/bin/g++
>> Fortran77 compiler: gfortran
>> Fortran77 compiler abs:
>> /usr/powerpc64-unknown-linux-gnu/gcc-bin/4.1.0/
>> gfortran
>> Fortran90 compiler: gfortran
>> Fortran90 compiler abs:
>> /usr/powerpc64-unknown-linux-gnu/gcc-bin/4.1.0/
>> gfortran
>> C profiling: yes
>> C++ profiling: yes
>> Fortran77 profiling: yes
>> Fortran90 profiling: yes
>> C++ exceptions: no
>> Thread support: posix (mpi: no,
>> progress: no)
>> Internal debug support: no
>> MPI parameter check: runtime
>> Memory profiling support: no
>> Memory debugging support: no
>> libltdl support: yes
>> MCA memory: ptmalloc2 (MCA v1.0, API
>> v1.0, Component v1.1.1)
>> MCA paffinity: linux (MCA v1.0, API v1.0,
>> Component v1.1.1)
>> MCA maffinity: first_use (MCA v1.0, API
>> v1.0, Component v1.1.1)
>> MCA timer: linux (MCA v1.0, API v1.0,
>> Component v1.1.1)
>> MCA allocator: basic (MCA v1.0, API v1.0,
>> Component v1.0)
>> MCA allocator: bucket (MCA v1.0, API v1.0,
>> Component v1.0)
>> MCA coll: basic (MCA v1.0, API v1.0,
>> Componentv1.1.1)
>>
>> MCA coll: hierarch (MCA v1.0, API v1.0,
>> Component v1.1.1)
>> MCA coll: self (MCA v1.0, API v1.0,
>> Component v1.1.1)
>> MCA coll: sm (MCA v1.0, API v1.0,
>> Component v1.1.1)
>> MCA coll: tuned (MCA v1.0, API v1.0,
>> Component v1.1.1)
>> MCA io: romio (MCA v1.0, API v1.0,
>> Component v1.1.1)
>> MCA mpool: gm (MCA v1.0, API v1.0,
>> Component v1.1.1)
>> MCA mpool: sm (MCA v1.0, API v1.0,
>> Component v1.1.1)
>> MCA pml: ob1 (MCA v1.0, API v1.0,
>> Component v1.1.1)
>> MCA bml: r2 (MCA v1.0, API v1.0,
>> Component v1.1.1)
>> MCA rcache: rb (MCA v1.0, API v1.0,
>> Component v1.1.1)
>> MCA btl: gm (MCA v1.0, API v1.0,
>> Component v1.1.1)
>> MCA btl: self (MCA v1.0, API v1.0,
>> Component v1.1.1)
>> MCA btl: sm (MCA v1.0, API v1.0,
>> Component v1.1.1)
>> MCA btl: tcp (MCA v1.0, API v1.0,
>> Component v1.0)
>> MCA topo: unity (MCA v1.0, API v1.0,
>> Component v1.1.1)
>> MCA osc: pt2pt (MCA v1.0, API v1.0,
>> Component v1.0)
>> MCA gpr: null (MCA v1.0, API v1.0,
>> Component v1.1.1)
>> MCA gpr: proxy (MCA v1.0, API v1.0,
>> Component v1.1.1)
>> MCA gpr: replica (MCA v1.0, API
>> v1.0,
>> Component v1.1.1)
>> MCA iof: proxy (MCA v1.0, API v1.0,
>> Component v1.1.1)
>> MCA iof: svc (MCA v1.0, API v1.0,
>> Component v1.1.1)
>> MCA ns: proxy (MCA v1.0, API v1.0,
>> Component v1.1.1)
>> MCA ns: replica (MCA v1.0, API
>> v1.0,
>> Component v1.1.1)
>> MCA oob: tcp (MCA v1.0, API v1.0,
>> Component v1.0)
>> MCA ras: dash_host (MCA v1.0, API
>> v1.0, Component v1.1.1)
>> MCA ras: hostfile (MCA v1.0, API
>> v1.0, Component v1.1.1)
>> MCA ras: localhost (MCA v1.0, API
>> v1.0, Component v1.1.1)
>> MCA ras: tm (MCA v1.0, API v1.0,
>> Component v1.1.1)
>> MCA rds: hostfile (MCA v1.0, API
>> v1.0, Component v1.1.1)
>> MCA rds: resfile (MCA v1.0, API v1.0,
>> Component v1.1.1)
>> MCA rmaps: round_robin (MCA v1.0, API
>> v1.0, Component v1.1.1)
>> MCA rmgr: proxy (MCA v1.0, API v1.0,
>> Component v1.1.1)
>> MCA rmgr: urm (MCA v1.0, API v1.0,
>> Component v1.1.1)
>> MCA rml: oob (MCA v1.0, API v1.0,
>> Component v1.1.1)
>> MCA pls: fork (MCA v1.0, API v1.0,
>> Component v1.1.1)
>> MCA pls: rsh (MCA v1.0, API v1.0,
>> Component v1.1.1)
>> MCA pls: tm (MCA v1.0, API v1.0,
>> Component v1.1.1)
>> MCA sds: env (MCA v1.0, API v1.0,
>> Component v1.1.1)
>> MCA sds: pipe (MCA v1.0, API v1.0,
>> Component v1.1.1)
>> MCA sds: seed (MCA v1.0, API v1.0,
>> Component v1.1.1)
>> MCA sds: singleton (MCA v1.0, API
>> v1.0, Component v1.1.1)
>> Configured as:
>> ./configure \
>> --prefix=$PREFIX \
>> --enable-mpi-f77 \
>> --enable-mpi-f90 \
>> --enable-mpi-profile \
>> --enable-mpi-cxx \
>> --enable-pty-support \
>> --enable-shared \
>> --enable-smp-locks \
>> --enable-io-romio \
>> --with-tm=/usr/local/pbs \
>> --without-xgrid \
>> --without-slurm \
>> --with-gm=/opt/gm
>>
>> Darwin node90.meldrew.clusters.umaine.edu
>> <http://node90.meldrew.clusters.umaine.edu> 8.6.0
>> Darwin Kernel Version 8.6.0: Tue Mar 7 16:58:48 PST
>> 2006; root:xnu-792.6.70.obj~1/RELEASE_PPC Power
>> Macintosh powerpc
>> node90:~/src/hpl jbronder$
>> /usr/local/ompi-xl/bin/ompi_info
>> Open MPI: 1.1.1a1r10670
>> Open MPI SVN revision: r10670
>> Open RTE: 1.1.1a1r10670
>> Open RTE SVN revision: r10670
>> OPAL: 1.1.1a1r10670
>> OPAL SVN revision: r10670
>> Prefix: /usr/local/ompi-xl
>> Configured architecture: powerpc-apple-darwin8.6.0
>> Configured by:
>> Configured on: Thu Jul 6 10:05:20 EDT
>> 2006
>> Configure host:
>> node90.meldrew.clusters.umaine.edu
>> <http://node90.meldrew.clusters.umaine.edu>
>> Built by: root
>> Built on: Thu Jul 6 10:37:40 EDT
>> 2006
>> Built host:
>> node90.meldrew.clusters.umaine.edu
>> <http://node90.meldrew.clusters.umaine.edu>
>> C bindings: yes
>> C++ bindings: yes
>> Fortran77 bindings: yes (lower case)
>> Fortran90 bindings: yes
>> Fortran90 bindings size: small
>> C compiler: /opt/ibmcmp/vac/6.0/bin/xlc
>> C compiler absolute: /opt/ibmcmp/vac/6.0/bin/xlc
>> C++ compiler: /opt/ibmcmp/vacpp/6.0/
>> bin/xlc++
>> C++ compiler absolute: /opt/ibmcmp/vacpp/6.0/
>> bin/xlc++
>> Fortran77 compiler: /opt/ibmcmp/xlf/8.1/bin/
>> xlf_r
>> Fortran77 compiler abs: /opt/ibmcmp/xlf/8.1/bin/
>> xlf_r
>> Fortran90 compiler: /opt/ibmcmp/xlf/8.1/bin/
>> xlf90_r
>> Fortran90 compiler abs: /opt/ibmcmp/xlf/8.1/bin/
>> xlf90_r
>> C profiling: yes
>> C++ profiling: yes
>> Fortran77 profiling: yes
>> Fortran90 profiling: yes
>> C++ exceptions: no
>> Thread support: posix (mpi: no,
>> progress: no)
>> Internal debug support: no
>> MPI parameter check: runtime
>> Memory profiling support: no
>> Memory debugging support: no
>> libltdl support: yes
>> MCA memory: darwin (MCA v1.0, API v1.0,
>> Component v1.1.1)
>> MCA maffinity: first_use (MCA v1.0, API
>> v1.0, Component v1.1.1)
>> MCA timer: darwin (MCA v1.0, API v1.0,
>> Component v1.1.1)
>> MCA allocator: basic (MCA v1.0, API v1.0,
>> Component v1.0)
>> MCA allocator: bucket (MCA v1.0, API v1.0,
>> Component v1.0)
>> MCA coll: basic (MCA v1.0, API v1.0,
>> Component v1.1.1)
>> MCA coll: hierarch (MCA v1.0, API
>> v1.0, Component v1.1.1)
>> MCA coll: self (MCA v1.0, API v1.0,
>> Component v1.1.1)
>> MCA coll: sm (MCA v1.0, API v1.0,
>> Component v1.1.1)
>> MCA coll: tuned (MCA v1.0, API v1.0,
>> Component v1.1.1)
>> MCA io: romio (MCA v1.0, API v1.0,
>> Component v1.1.1)
>> MCA mpool: sm (MCA v1.0, API v1.0,
>> Component v1.1.1)
>> MCA mpool: gm (MCA v1.0, API v1.0,
>> Component v1.1.1)
>> MCA pml: ob1 (MCA v1.0, API v1.0,
>> Component v1.1.1)
>> MCA bml: r2 (MCA v1.0, API v1.0,
>> Component v1.1.1)
>> MCA rcache: rb (MCA v1.0, API v1.0,
>> Component v1.1.1)
>> MCA btl: self (MCA v1.0, API v1.0,
>> Component v1.1.1)
>> MCA btl: sm (MCA v1.0, API v1.0,
>> Component v1.1.1)
>> MCA btl: gm (MCA v1.0, API v1.0,
>> Component v1.1.1)
>> MCA btl: tcp (MCA v1.0, API v1.0,
>> Component v1.0)
>> MCA topo: unity (MCA v1.0, API v1.0,
>> Component v1.1.1)
>> MCA osc: pt2pt (MCA v1.0, API v1.0,
>> Component v1.0)
>> MCA gpr: null (MCA v1.0, API v1.0,
>> Component v1.1.1)
>> MCA gpr: proxy (MCA v1.0, API v1.0,
>> Component v1.1.1)
>> MCA gpr: replica (MCA v1.0, API
>> v1.0,
>> Component v1.1.1)
>> MCA iof: proxy (MCA v1.0, API v1.0,
>> Component v1.1.1)
>> MCA iof: svc (MCA v1.0, API v1.0,
>> Component v1.1.1)
>> MCA ns: proxy (MCA v1.0, API v1.0,
>> Component v1.1.1)
>> MCA ns: replica (MCA v1.0, API
>> v1.0,
>> Component v1.1.1)
>> MCA oob: tcp (MCA v1.0, API v1.0,
>> Component v1.0)
>> MCA ras: dash_host (MCA v1.0, API
>> v1.0, Component v1.1.1)
>> MCA ras: hostfile (MCA v1.0, API
>> v1.0, Component v1.1.1)
>> MCA ras: localhost (MCA v1.0, API
>> v1.0, Component v1.1.1)
>> MCA ras: tm (MCA v1.0, API v1.0,
>> Component v1.1.1)
>> MCA rds: hostfile (MCA v1.0, API v1.0,
>> Component v1.1.1)
>> MCA rds: resfile (MCA v1.0, API
>> v1.0,
>> Component v1.1.1)
>> MCA rmaps: round_robin (MCA v1.0, API
>> v1.0, Component v1.1.1)
>> MCA rmgr: proxy (MCA v1.0, API v1.0,
>> Component v1.1.1)
>> MCA rmgr: urm (MCA v1.0, API v1.0,
>> Component v1.1.1)
>> MCA rml: oob (MCA v1.0, API v1.0,
>> Component v1.1.1)
>> MCA pls: fork (MCA v1.0, API v1.0,
>> Component v1.1.1)
>> MCA pls: rsh (MCA v1.0, API v1.0,
>> Component v1.1.1)
>> MCA pls: tm (MCA v1.0, API v1.0,
>> Component v1.1.1)
>> MCA sds: env (MCA v1.0, API v1.0,
>> Component v1.1.1)
>> MCA sds: seed (MCA v1.0, API v1.0,
>> Component v1.1.1)
>> MCA sds: singleton (MCA v1.0, API
>> v1.0, Component v1.1.1)
>> MCA sds: pipe (MCA v1.0, API v1.0,
>> Component v1.1.1)
>> Configured as:
>> ./configure \
>> --prefix=$PREFIX \
>> --with-tm=/usr/local/pbs/ \
>> --with-gm=/opt/gm \
>> --enable-static \
>> --disable-cxx
>> On 7/3/06, *George Bosilca* < bosilca_at_[hidden]
>> <mailto:bosilca_at_[hidden]>> wrote:
>> Bernard,
>>
>> A bug in the Open MPI GM driver was discovered after
>> the 1.1 release.
>> A patch for the 1.1 is on the way. However, I don't
>> know if it will
>> be available before the 1.1.1. Meanwhile, you can use
>> the nightly
>> build version or a fresh check-out from the SVN
>> repository. Both of
>> them have the GM bug corrected.
>>
>> Sorry for the troubles,
>> george.
>>
>> On Jul 3, 2006, at 12:58 PM, Borenstein, Bernard S
>> wrote:
>>
>>> I've built and sucessfully run the Nasa Overflow
>> 2.0aa program with
>>> Openmpi 1.0.2. I'm running on an opteron linux
>> cluster running SLES 9
>>> and GM 2.0.24. I built Openmpi 1.1 with the intel 9
>> compilers and
>>> try to
>>> run Overflow 2.0aa with myrinet, it get what looks
>> like a data
>>> corruption error and the program dies quickly.
>>> There are no mpi errors at all.If I run using GIGE
>> (--mca btl
>>> self,tcp),
>>> the program runs to competion correctly. Here is my
>> ompi_info
>>> output :
>>>
>>> bsb3227_at_mahler:~/openmpi_1.1/bin> ./ompi_info
>>> Open MPI: 1.1
>>> Open MPI SVN revision: r10477
>>> Open RTE: 1.1
>>> Open RTE SVN revision: r10477
>>> OPAL: 1.1
>>> OPAL SVN revision: r10477
>>> Prefix: /home/bsb3227/openmpi_1.1
>>> Configured architecture: x86_64-unknown-linux-gnu
>>> Configured by: bsb3227
>>> Configured on: Fri Jun 30 07:08:54 PDT 2006
>>> Configure host: mahler
>>> Built by: bsb3227
>>> Built on: Fri Jun 30 07:54:46 PDT 2006
>>> Built host: mahler
>>> C bindings: yes
>>> C++ bindings: yes
>>> Fortran77 bindings: yes (all)
>>> Fortran90 bindings: yes
>>> Fortran90 bindings size: small
>>> C compiler: icc
>>> C compiler absolute: /opt/intel/cce/9.0.25/bin/icc
>>> C++ compiler: icpc
>>> C++ compiler absolute: /opt/intel/cce/9.0.25/bin/icpc
>>> Fortran77 compiler: ifort
>>> Fortran77 compiler abs: /opt/intel/fce/9.0.25/bin/ifort
>>> Fortran90 compiler: /opt/intel/fce/9.0.25/bin/ifort
>>> Fortran90 compiler abs:
>> /opt/intel/fce/9.0.25/bin/ifort
>>> C profiling: yes
>>> C++ profiling: yes
>>> Fortran77 profiling: yes
>>> Fortran90 profiling: yes
>>> C++ exceptions: no
>>> Thread support: posix (mpi: no, progress: no)
>>> Internal debug support: no
>>> MPI parameter check: runtime
>>> Memory profiling support: no
>>> Memory debugging support: no
>>> libltdl support: yes
>>> MCA memory: ptmalloc2 (MCA v1.0, API
>> v1.0, Component
>>> v1.1)
>>> MCA paffinity: linux (MCA v1.0, API v1.0,
>> Component v1.1)
>>> MCA maffinity: first_use (MCA v1.0, API
>> v1.0, Component
>>> v1.1)
>>> MCA maffinity: libnuma (MCA v1.0, API
>> v1.0, Component v1.1)
>>> MCA timer: linux (MCA v1.0, API v1.0,
>> Component v1.1)
>>> MCA allocator: basic (MCA v1.0, API v1.0,
>> Component v1.0)
>>> MCA allocator: bucket (MCA v1.0, API v1.0,
>> Component v1.0)
>>> MCA coll: basic (MCA v1.0, API v1.0,
>> Component v1.1)
>>> MCA coll: hierarch (MCA v1.0, API
>> v1.0, Component
>>> v1.1)
>>> MCA coll: self (MCA v1.0, API v1.0,
>> Component v1.1)
>>> MCA coll: sm (MCA v1.0, API v1.0,
>> Component v1.1)
>>> MCA coll: tuned (MCA v1.0, API v1.0,
>> Component v1.1)
>>> MCA io: romio (MCA v1.0, API v1.0,
>> Component v1.1)
>>> MCA mpool: sm (MCA v1.0, API v1.0,
>> Component v1.1)
>>> MCA mpool: gm (MCA v1.0, API v1.0,
>> Component v1.1)
>>> MCA pml: ob1 (MCA v1.0, API v1.0,
>> Component v1.1)
>>> MCA bml: r2 (MCA v1.0, API v1.0,
>> Component v1.1)
>>> MCA rcache: rb (MCA v1.0, API v1.0,
>> Component v1.1)
>>> MCA btl: self (MCA v1.0, API v1.0,
>> Component v1.1)
>>> MCA btl: sm (MCA v1.0, API v1.0,
>> Component v1.1)
>>> MCA btl: gm (MCA v1.0, API v1.0,
>> Component v1.1)
>>> MCA btl: tcp (MCA v1.0, API v1.0,
>> Component v1.0)
>>> MCA topo: unity (MCA v1.0, API v1.0,
>> Component v1.1)
>>> MCA osc: pt2pt (MCA v1.0, API v1.0,
>> Component v1.0)
>>> MCA gpr: null (MCA v1.0, API v1.0,
>> Component v1.1)
>>> MCA gpr: proxy (MCA v1.0, API v1.0,
>> Component v1.1)
>>> MCA gpr: replica (MCA v1.0, API
>> v1.0, Component v1.1)
>>> MCA iof: proxy (MCA v1.0, API v1.0,
>> Component v1.1)
>>> MCA iof: svc (MCA v1.0, API v1.0,
>> Component v1.1)
>>> MCA ns: proxy (MCA v1.0, API v1.0,
>> Component v1.1)
>>> MCA ns: replica (MCA v1.0, API
>> v1.0, Component v1.1)
>>> MCA oob: tcp (MCA v1.0, API v1.0,
>> Component v1.0)
>>> MCA ras: dash_host (MCA v1.0, API
>> v1.0, Component
>>> v1.1)
>>> MCA ras: hostfile (MCA v1.0, API
>> v1.0, Component
>>> v1.1)
>>> MCA ras: localhost (MCA v1.0, API
>> v1.0, Component
>>> v1.1)
>>> MCA ras: slurm (MCA v1.0, API v1.0,
>> Component v1.1)
>>> MCA ras: tm (MCA v1.0, API v1.0,
>> Component v1.1)
>>> MCA rds: hostfile (MCA v1.0, API
>> v1.0, Component
>>> v1.1)
>>> MCA rds: resfile (MCA v1.0, API
>> v1.0, Component v1.1)
>>> MCA rmaps: round_robin (MCA v1.0, API
>> v1.0, Component
>>> v1.1)
>>> MCA rmgr: proxy (MCA v1.0, API v1.0,
>> Component v1.1)
>>> MCA rmgr: urm (MCA v1.0, API v1.0,
>> Component v1.1)
>>> MCA rml: oob (MCA v1.0, API v1.0,
>> Component v1.1)
>>> MCA pls: fork (MCA v1.0, API v1.0,
>> Component v1.1)
>>> MCA pls: rsh (MCA v1.0, API v1.0,
>> Component v1.1)
>>> MCA pls: slurm (MCA v1.0, API v1.0,
>> Component v1.1)
>>> MCA pls: tm (MCA v1.0, API v1.0,
>> Component v1.1)
>>> MCA sds: env (MCA v1.0, API v1.0,
>> Component v1.1)
>>> MCA sds: seed (MCA v1.0, APIv1.0,
>> Component v1.1)
>>
>>> MCA sds: singleton (MCA v1.0, API
>> v1.0, Component
>>> v1.1)
>>> MCA sds: pipe (MCAv1.0, API v1.0,
>> Component v1.1)
>>
>>> MCA sds: slurm (MCA v1.0, API v1.0,
>> Component v1.1)
>>>
>>> Here is the ifconfig for one of the nodes :
>>>
>>> bsb3227_at_m045:~> /sbin/ifconfig
>>> eth0 Link encap:Ethernet HWaddr 00:50:45:5D:CD:FE
>>> inet addr:10.241.194.45
>> <http://10.241.194.45> Bcast: 10.241.195.255
>> <http://10.241.195.255>
>>> Mask:255.255.254.0 <http://255.255.254.0>
>>> inet6 addr: fe80::250:45ff:fe5d:cdfe/64
>> Scope:Link
>>> UP BROADCAST NOTRAILERS RUNNING
>> MULTICAST MTU:1500
>>> Metric:1
>>> RX packets:39913407 errors:0 dropped:0
>> overruns:0 frame:0
>>> TX packets:48794587 errors:0 dropped:0
>> overruns:0 carrier:0
>>> collisions:0 txqueuelen:1000
>>> RX bytes:31847343907 (30371.9 Mb) TX
>> bytes:48231713866
>>> (45997.3 Mb)
>>> Interrupt:19
>>>
>>> eth1 Link encap:Ethernet HWaddr 00:50:45:5D:CD:FF
>>> inet6 addr: fe80::250:45ff:fe5d:cdff/64
>> Scope:Link
>>> UP BROADCAST MULTICAST MTU:1500 Metric:1
>>> RX packets:0 errors:0 dropped:0 overruns:0
>> frame:0
>>> TX packets:0 errors:0 dropped:0 overruns:0
>> carrier:0
>>> collisions:0 txqueuelen:1000
>>> RX bytes:0 (0.0 b) TX bytes:0 (0.0 b)
>>> Interrupt:19
>>>
>>> lo Link encap:Local Loopback
>>> inet addr: 127.0.0.1
>> <http://127.0.0.1> Mask: 255.0.0.0 <http://
>> 255.0.0.0>
>>> inet6 addr: ::1/128 Scope:Host
>>> UP LOOPBACK RUNNING MTU:16436 Metric:1
>>> RX packets:23141 errors:0 dropped:0
>> overruns:0 frame:0
>>> TX packets:23141 errors:0 dropped:0
>> overruns:0 carrier:0
>>> collisions:0 txqueuelen:0
>>> RX bytes:20145689 (19.2 Mb) TX
>> bytes:20145689 (19.2 Mb)
>>>
>>> I hope someone can give me some guidance on how to
>> debug this problem.
>>> Thanx in advance for any help
>>> that can be provided.
>>>
>>> Bernie Borenstein
>>> The Boeing Company
>>> <config.log.gz>
>>> _______________________________________________
>>> users mailing list
>>> users_at_[hidden] <mailto:users_at_[hidden]>
>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>
>> "Half of what I say is meaningless; but I say it so
>> that the other
>> half may reach you"
>> Kahlil Gibran
>>
>>
>> _______________________________________________
>> users mailing list
>> users_at_[hidden] <mailto:users_at_[hidden]>
>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>
>>
>> _______________________________________________
>> users mailing list
>> users_at_[hidden] <mailto:users_at_[hidden]>
>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>
>>
>> _______________________________________________
>> users mailing list
>> users_at_[hidden] <mailto:users_at_[hidden]>
>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>> <http://www.open-mpi.org/mailman/listinfo.cgi/users>
>>
>>
>> ---------------------------------------------------------------------
>> ---
>>
>> _______________________________________________
>> users mailing list
>> users_at_[hidden]
>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>
> --
> Justin Bronder
> University of Maine, Orono
>
> Advanced Computing Research Lab
> 20 Godfrey Dr
> Orono, ME 04473
> www.clusters.umaine.edu
>
> Mathematics Department
> 425 Neville Hall
> Orono, ME 04469
>
>
> _______________________________________________
> users mailing list
> users_at_[hidden]
> http://www.open-mpi.org/mailman/listinfo.cgi/users