Open MPI logo

Open MPI User's Mailing List Archives

  |   Home   |   Support   |   FAQ   |   all Open MPI User's mailing list

From: Jeff Squyres (jsquyres_at_[hidden])
Date: 2005-11-10 13:24:54


One minor thing that I notice in your ompi_info output -- your build
and run machines are different (kfc vs. clement).

Are these both FC4 machines, or are they different OS's/distros?

On Nov 10, 2005, at 10:01 AM, Clement Chu wrote:

> [clement_at_kfc TestMPI]$ mpirun -d -np 2 test
> [kfc:29199] procdir: (null)
> [kfc:29199] jobdir: (null)
> [kfc:29199] unidir:
> /tmp/openmpi-sessions-clement_at_kfc_0/default-universe
> [kfc:29199] top: openmpi-sessions-clement_at_kfc_0
> [kfc:29199] tmp: /tmp
> [kfc:29199] [0,0,0] setting up session dir with
> [kfc:29199] tmpdir /tmp
> [kfc:29199] universe default-universe-29199
> [kfc:29199] user clement
> [kfc:29199] host kfc
> [kfc:29199] jobid 0
> [kfc:29199] procid 0
> [kfc:29199] procdir:
> /tmp/openmpi-sessions-clement_at_kfc_0/default-universe-29199/0/0
> [kfc:29199] jobdir:
> /tmp/openmpi-sessions-clement_at_kfc_0/default-universe-29199/0
> [kfc:29199] unidir:
> /tmp/openmpi-sessions-clement_at_kfc_0/default-universe-29199
> [kfc:29199] top: openmpi-sessions-clement_at_kfc_0
> [kfc:29199] tmp: /tmp
> [kfc:29199] [0,0,0] contact_file
> /tmp/openmpi-sessions-clement_at_kfc_0/default-universe-29199/universe-
> setup.txt
> [kfc:29199] [0,0,0] wrote setup file
> [kfc:29199] pls:rsh: local csh: 0, local bash: 1
> [kfc:29199] pls:rsh: assuming same remote shell as local shell
> [kfc:29199] pls:rsh: remote csh: 0, remote bash: 1
> [kfc:29199] pls:rsh: final template argv:
> [kfc:29199] pls:rsh: ssh <template> orted --debug --bootproxy 1
> --name <template> --num_procs 2 --vpid_start 0 --nodename <template>
> --universe clement_at_kfc:default-universe-29199 --nsreplica
> "0.0.0;tcp://192.168.11.101:32784" --gprreplica
> "0.0.0;tcp://192.168.11.101:32784" --mpi-call-yield 0
> [kfc:29199] pls:rsh: launching on node localhost
> [kfc:29199] pls:rsh: oversubscribed -- setting mpi_yield_when_idle to 1
> (1 2)
> [kfc:29199] sess_dir_finalize: proc session dir not empty - leaving
> [kfc:29199] spawn: in job_state_callback(jobid = 1, state = 0xa)
> mpirun noticed that job rank 1 with PID 0 on node "localhost" exited on
> signal 11.
> [kfc:29199] sess_dir_finalize: proc session dir not empty - leaving
> [kfc:29199] spawn: in job_state_callback(jobid = 1, state = 0x9)
> [kfc:29199] ERROR: A daemon on node localhost failed to start as
> expected.
> [kfc:29199] ERROR: There may be more information available from
> [kfc:29199] ERROR: the remote shell (see above).
> [kfc:29199] The daemon received a signal 11.
> 1 additional process aborted (not shown)
> [kfc:29199] sess_dir_finalize: found proc session dir empty - deleting
> [kfc:29199] sess_dir_finalize: found job session dir empty - deleting
> [kfc:29199] sess_dir_finalize: found univ session dir empty - deleting
> [kfc:29199] sess_dir_finalize: top session dir not empty - leaving
>
>
> opmi_info output message:
>
> [clement_at_kfc TestMPI]$ ompi_info
> Open MPI: 1.0rc5r8053
> Open MPI SVN revision: r8053
> Open RTE: 1.0rc5r8053
> Open RTE SVN revision: r8053
> OPAL: 1.0rc5r8053
> OPAL SVN revision: r8053
> Prefix: /home/clement/openmpi
> Configured architecture: i686-pc-linux-gnu
> Configured by: clement
> Configured on: Fri Nov 11 00:37:23 EST 2005
> Configure host: kfc
> Built by: clement
> Built on: Fri Nov 11 00:59:26 EST 2005
> Built host: kfc
> C bindings: yes
> C++ bindings: yes
> Fortran77 bindings: yes (all)
> Fortran90 bindings: yes
> C compiler: gcc
> C compiler absolute: /usr/bin/gcc
> C++ compiler: g++
> C++ compiler absolute: /usr/bin/g++
> Fortran77 compiler: gfortran
> Fortran77 compiler abs: /usr/bin/gfortran
> Fortran90 compiler: gfortran
> Fortran90 compiler abs: /usr/bin/gfortran
> C profiling: yes
> C++ profiling: yes
> Fortran77 profiling: yes
> Fortran90 profiling: yes
> C++ exceptions: no
> Thread support: posix (mpi: no, progress: no)
> Internal debug support: no
> MPI parameter check: runtime
> Memory profiling support: no
> Memory debugging support: no
> libltdl support: 1
> MCA memory: malloc_hooks (MCA v1.0, API v1.0, Component
> v1.0)
> MCA paffinity: linux (MCA v1.0, API v1.0, Component v1.0)
> MCA maffinity: first_use (MCA v1.0, API v1.0, Component
> v1.0)
> MCA timer: linux (MCA v1.0, API v1.0, Component v1.0)
> MCA allocator: basic (MCA v1.0, API v1.0, Component v1.0)
> MCA allocator: bucket (MCA v1.0, API v1.0, Component v1.0)
> MCA coll: basic (MCA v1.0, API v1.0, Component v1.0)
> MCA coll: self (MCA v1.0, API v1.0, Component v1.0)
> MCA coll: sm (MCA v1.0, API v1.0, Component v1.0)
> MCA io: romio (MCA v1.0, API v1.0, Component v1.0)
> MCA mpool: sm (MCA v1.0, API v1.0, Component v1.0)
> MCA pml: ob1 (MCA v1.0, API v1.0, Component v1.0)
> MCA pml: teg (MCA v1.0, API v1.0, Component v1.0)
> MCA pml: uniq (MCA v1.0, API v1.0, Component v1.0)
> MCA ptl: self (MCA v1.0, API v1.0, Component v1.0)
> MCA ptl: sm (MCA v1.0, API v1.0, Component v1.0)
> MCA ptl: tcp (MCA v1.0, API v1.0, Component v1.0)
> MCA btl: self (MCA v1.0, API v1.0, Component v1.0)
> MCA btl: sm (MCA v1.0, API v1.0, Component v1.0)
> MCA btl: tcp (MCA v1.0, API v1.0, Component v1.0)
> MCA topo: unity (MCA v1.0, API v1.0, Component v1.0)
> MCA gpr: null (MCA v1.0, API v1.0, Component v1.0)
> MCA gpr: proxy (MCA v1.0, API v1.0, Component v1.0)
> MCA gpr: replica (MCA v1.0, API v1.0, Component v1.0)
> MCA iof: proxy (MCA v1.0, API v1.0, Component v1.0)
> MCA iof: svc (MCA v1.0, API v1.0, Component v1.0)
> MCA ns: proxy (MCA v1.0, API v1.0, Component v1.0)
> MCA ns: replica (MCA v1.0, API v1.0, Component v1.0)
> MCA oob: tcp (MCA v1.0, API v1.0, Component v1.0)
> MCA ras: dash_host (MCA v1.0, API v1.0, Component
> v1.0)
> MCA ras: hostfile (MCA v1.0, API v1.0, Component v1.0)
> MCA ras: localhost (MCA v1.0, API v1.0, Component
> v1.0)
> MCA ras: slurm (MCA v1.0, API v1.0, Component v1.0)
> MCA rds: hostfile (MCA v1.0, API v1.0, Component v1.0)
> MCA rds: resfile (MCA v1.0, API v1.0, Component v1.0)
> MCA rmaps: round_robin (MCA v1.0, API v1.0, Component
> v1.0)
> MCA rmgr: proxy (MCA v1.0, API v1.0, Component v1.0)
> MCA rmgr: urm (MCA v1.0, API v1.0, Component v1.0)
> MCA rml: oob (MCA v1.0, API v1.0, Component v1.0)
> MCA pls: fork (MCA v1.0, API v1.0, Component v1.0)
> MCA pls: proxy (MCA v1.0, API v1.0, Component v1.0)
> MCA pls: rsh (MCA v1.0, API v1.0, Component v1.0)
> MCA pls: slurm (MCA v1.0, API v1.0, Component v1.0)
> MCA sds: env (MCA v1.0, API v1.0, Component v1.0)
> MCA sds: pipe (MCA v1.0, API v1.0, Component v1.0)
> MCA sds: seed (MCA v1.0, API v1.0, Component v1.0)
> MCA sds: singleton (MCA v1.0, API v1.0, Component
> v1.0)
> MCA sds: slurm (MCA v1.0, API v1.0, Component v1.0)
> [clement_at_kfc TestMPI]$

-- 
{+} Jeff Squyres
{+} The Open MPI Project
{+} http://www.open-mpi.org/