Open MPI logo

Open MPI User's Mailing List Archives

  |   Home   |   Support   |   FAQ   |   all Open MPI User's mailing list

Subject: Re: [OMPI users] cannot build openmpi-1.9r28290 on Linux/Solaris
From: Ralph Castain (rhc_at_[hidden])
Date: 2013-04-05 13:44:50


On Apr 5, 2013, at 12:33 AM, Siegmar Gross <Siegmar.Gross_at_[hidden]> wrote:

> Hi
>
> today I tried to install openmpi-1.9r28290 and got the following errors.
>
> Solaris 10, Sparc, Sun C 5.12, 32-bit version of openmpi
> Solaris 10, x86_64, Sun C 5.12, 32-bit version of openmpi
> Solaris 10, Sparc, Sun C 5.12, 64-bit version of openmpi
> Solaris 10, x86_64, Sun C 5.12, 64-bit version of openmpi
> ---------------------------------------------------------
>
> ...
> CC topology-solaris.lo
> "../../../../../../../openmpi-1.9r28290/opal/mca/hwloc/hwloc152/hwloc/src/topolo
> gy-solaris.c", line 226: undefined symbol: binding
> "../../../../../../../openmpi-1.9r28290/opal/mca/hwloc/hwloc152/hwloc/src/topolo
> gy-solaris.c", line 227: undefined symbol: hwloc_set
> "../../../../../../../openmpi-1.9r28290/opal/mca/hwloc/hwloc152/hwloc/src/topolo
> gy-solaris.c", line 227: warning: improper pointer/integer combination: arg #1
> cc: acomp failed for ../../../../../../../openmpi-1.9r28290/opal/mca/hwloc/hwloc
> 152/hwloc/src/topology-solaris.c
> make[4]: *** [topology-solaris.lo] Error 1
> ...
>

Found a missing variable declaration - try with r28293 or above.

>
>
>
> openSuSE Linux 12.1, x86_64, Sun C 5.12, 32-bit version of openmpi
> openSuSE Linux 12.1, x86_64, Sun C 5.12, 64-bit version of openmpi
> ------------------------------------------------------------------
>
> ...
> PPFC mpi-f08-sizeof.lo
> PPFC mpi-f08.lo
> "../../../../../openmpi-1.9r28290/ompi/mpi/fortran/use-mpi-f08/mpi-f08.F90", Lin
> e = 1, Column = 1: INTERNAL: Interrupt: Segmentation fault
> make[2]: *** [mpi-f08.lo] Error 1
> make[2]: Leaving directory `/export2/src/openmpi-1.9/openmpi-1.9-Linux.x86_64.32
> _cc/ompi/mpi/fortran/use-mpi-f08'
> make[1]: *** [all-recursive] Error 1
> ...
>

I have to defer the Fortran stuff to Jeff.

>
> I could built an older version.
>
> Package: Open MPI root_at_linpc1 Distribution
> Open MPI: 1.9r28209
> Open MPI repo revision: r28209
> Open MPI release date: Mar 25, 2013 (nightly snapshot tarball)
> Open RTE: 1.9
> Open RTE repo revision: r28134M
> Open RTE release date: Feb 28, 2013
> OPAL: 1.9
> OPAL repo revision: r28134M
> OPAL release date: Feb 28, 2013
> MPI API: 2.1
> Ident string: 1.9r28209
> Prefix: /usr/local/ompi-java_64_cc
> Configured architecture: x86_64-unknown-linux-gnu
> Configure host: linpc1
> Configured by: root
> Configured on: Tue Mar 26 15:54:59 CET 2013
> Configure host: linpc1
> Built by: root
> Built on: Tue Mar 26 16:31:01 CET 2013
> Built host: linpc1
> C bindings: yes
> C++ bindings: yes
> Fort mpif.h: yes (all)
> Fort use mpi: yes (full: ignore TKR)
> Fort use mpi size: deprecated-ompi-info-value
> Fort use mpi_f08: yes
> Fort mpi_f08 compliance: The mpi_f08 module is available, but due to
> limitations in the f95 compiler, does not support the following: array
> subsections, ABSTRACT INTERFACE function pointers, Fortran '08-specified
> ASYNCHRONOUS behavior, PROCEDUREs, direct passthru (where possible) to
> underlying Open MPI's C functionality
> Fort mpi_f08 subarrays: no
> Java bindings: yes
> C compiler: cc
> C compiler absolute: /opt/solstudio12.3/bin/cc
> C compiler family name: SUN
> C compiler version: 0x5120
> C++ compiler: CC
> C++ compiler absolute: /opt/solstudio12.3/bin/CC
> Fort compiler: f95
> Fort compiler abs: /opt/solstudio12.3/bin/f95
> Fort ignore TKR: yes (!$PRAGMA IGNORE_TKR)
> Fort 08 assumed shape: no
> Fort optional args: yes
> Fort BIND(C): yes
> Fort PRIVATE: yes
> Fort ABSTRACT: no
> Fort ASYNCHRONOUS: no
> Fort PROCEDURE: no
> Fort f08 using wrappers: yes
> C profiling: yes
> C++ profiling: yes
> Fort mpif.h profiling: yes
> Fort use mpi profiling: yes
> Fort use mpi_f08 prof: yes
> C++ exceptions: yes
> Thread support: posix (MPI_THREAD_MULTIPLE: yes, OPAL support: yes,
> OMPI progress: no, ORTE progress: no, Event lib: no)
> Sparse Groups: no
> Internal debug support: yes
> MPI interface warnings: yes
> MPI parameter check: runtime
> Memory profiling support: no
> Memory debugging support: no
> libltdl support: yes
> Heterogeneous support: yes
> mpirun default --prefix: no
> MPI I/O support: yes
> MPI_WTIME support: gettimeofday
> Symbol vis. support: yes
> Host topology support: yes
> MPI extensions:
> FT Checkpoint support: no (checkpoint thread: no)
> C/R Enabled Debugging: no
> VampirTrace support: yes
> MPI_MAX_PROCESSOR_NAME: 256
> MPI_MAX_ERROR_STRING: 256
> MPI_MAX_OBJECT_NAME: 64
> MPI_MAX_INFO_KEY: 36
> MPI_MAX_INFO_VAL: 256
> MPI_MAX_PORT_NAME: 1024
> MPI_MAX_DATAREP_STRING: 128
> MCA backtrace: execinfo (MCA v2.0, API v2.0, Component v1.9)
> MCA event: libevent2019 (MCA v2.0, API v2.0, Component v1.9)
> MCA hwloc: hwloc152 (MCA v2.0, API v2.0, Component v1.9)
> MCA if: linux_ipv6 (MCA v2.0, API v2.0, Component v1.9)
> MCA if: posix_ipv4 (MCA v2.0, API v2.0, Component v1.9)
> MCA installdirs: env (MCA v2.0, API v2.0, Component v1.9)
> MCA installdirs: config (MCA v2.0, API v2.0, Component v1.9)
> MCA memory: linux (MCA v2.0, API v2.0, Component v1.9)
> MCA shmem: mmap (MCA v2.0, API v2.0, Component v1.9)
> MCA shmem: posix (MCA v2.0, API v2.0, Component v1.9)
> MCA shmem: sysv (MCA v2.0, API v2.0, Component v1.9)
> MCA timer: linux (MCA v2.0, API v2.0, Component v1.9)
> MCA dfs: app (MCA v2.0, API v1.0, Component v1.9)
> MCA dfs: orted (MCA v2.0, API v1.0, Component v1.9)
> MCA dfs: test (MCA v2.0, API v1.0, Component v1.9)
> MCA errmgr: default_app (MCA v2.0, API v3.0, Component v1.9)
> MCA errmgr: default_hnp (MCA v2.0, API v3.0, Component v1.9)
> MCA errmgr: default_orted (MCA v2.0, API v3.0, Component v1.9)
> MCA ess: env (MCA v2.0, API v3.0, Component v1.9)
> MCA ess: hnp (MCA v2.0, API v3.0, Component v1.9)
> MCA ess: singleton (MCA v2.0, API v3.0, Component v1.9)
> MCA ess: slurm (MCA v2.0, API v3.0, Component v1.9)
> MCA ess: tool (MCA v2.0, API v3.0, Component v1.9)
> MCA filem: raw (MCA v2.0, API v2.0, Component v1.9)
> MCA filem: rsh (MCA v2.0, API v2.0, Component v1.9)
> MCA grpcomm: bad (MCA v2.0, API v2.0, Component v1.9)
> MCA iof: hnp (MCA v2.0, API v2.0, Component v1.9)
> MCA iof: mr_hnp (MCA v2.0, API v2.0, Component v1.9)
> MCA iof: mr_orted (MCA v2.0, API v2.0, Component v1.9)
> MCA iof: orted (MCA v2.0, API v2.0, Component v1.9)
> MCA iof: tool (MCA v2.0, API v2.0, Component v1.9)
> MCA odls: default (MCA v2.0, API v2.0, Component v1.9)
> MCA oob: tcp (MCA v2.0, API v2.0, Component v1.9)
> MCA plm: rsh (MCA v2.0, API v2.0, Component v1.9)
> MCA plm: slurm (MCA v2.0, API v2.0, Component v1.9)
> MCA ras: loadleveler (MCA v2.0, API v2.0, Component v1.9)
> MCA ras: simulator (MCA v2.0, API v2.0, Component v1.9)
> MCA ras: slurm (MCA v2.0, API v2.0, Component v1.9)
> MCA rmaps: lama (MCA v2.0, API v2.0, Component v1.9)
> MCA rmaps: ppr (MCA v2.0, API v2.0, Component v1.9)
> MCA rmaps: rank_file (MCA v2.0, API v2.0, Component v1.9)
> MCA rmaps: resilient (MCA v2.0, API v2.0, Component v1.9)
> MCA rmaps: round_robin (MCA v2.0, API v2.0, Component v1.9)
> MCA rmaps: seq (MCA v2.0, API v2.0, Component v1.9)
> MCA rmaps: staged (MCA v2.0, API v2.0, Component v1.9)
> MCA rml: oob (MCA v2.0, API v2.0, Component v1.9)
> MCA routed: binomial (MCA v2.0, API v2.0, Component v1.9)
> MCA routed: debruijn (MCA v2.0, API v2.0, Component v1.9)
> MCA routed: direct (MCA v2.0, API v2.0, Component v1.9)
> MCA routed: radix (MCA v2.0, API v2.0, Component v1.9)
> MCA state: app (MCA v2.0, API v1.0, Component v1.9)
> MCA state: hnp (MCA v2.0, API v1.0, Component v1.9)
> MCA state: novm (MCA v2.0, API v1.0, Component v1.9)
> MCA state: orted (MCA v2.0, API v1.0, Component v1.9)
> MCA state: staged_hnp (MCA v2.0, API v1.0, Component v1.9)
> MCA state: staged_orted (MCA v2.0, API v1.0, Component v1.9)
> MCA allocator: basic (MCA v2.0, API v2.0, Component v1.9)
> MCA allocator: bucket (MCA v2.0, API v2.0, Component v1.9)
> MCA bml: r2 (MCA v2.0, API v2.0, Component v1.9)
> MCA btl: self (MCA v2.0, API v2.0, Component v1.9)
> MCA btl: sm (MCA v2.0, API v2.0, Component v1.9)
> MCA btl: tcp (MCA v2.0, API v2.0, Component v1.9)
> MCA coll: basic (MCA v2.0, API v2.0, Component v1.9)
> MCA coll: hierarch (MCA v2.0, API v2.0, Component v1.9)
> MCA coll: inter (MCA v2.0, API v2.0, Component v1.9)
> MCA coll: libnbc (MCA v2.0, API v2.0, Component v1.9)
> MCA coll: self (MCA v2.0, API v2.0, Component v1.9)
> MCA coll: sm (MCA v2.0, API v2.0, Component v1.9)
> MCA coll: tuned (MCA v2.0, API v2.0, Component v1.9)
> MCA fbtl: posix (MCA v2.0, API v2.0, Component v1.9)
> MCA fcoll: dynamic (MCA v2.0, API v2.0, Component v1.9)
> MCA fcoll: individual (MCA v2.0, API v2.0, Component v1.9)
> MCA fcoll: static (MCA v2.0, API v2.0, Component v1.9)
> MCA fcoll: two_phase (MCA v2.0, API v2.0, Component v1.9)
> MCA fcoll: ylib (MCA v2.0, API v2.0, Component v1.9)
> MCA fs: ufs (MCA v2.0, API v2.0, Component v1.9)
> MCA io: ompio (MCA v2.0, API v2.0, Component v1.9)
> MCA io: romio (MCA v2.0, API v2.0, Component v1.9)
> MCA mpool: grdma (MCA v2.0, API v2.0, Component v1.9)
> MCA mpool: sm (MCA v2.0, API v2.0, Component v1.9)
> MCA osc: pt2pt (MCA v2.0, API v2.0, Component v1.9)
> MCA osc: rdma (MCA v2.0, API v2.0, Component v1.9)
> MCA pml: v (MCA v2.0, API v2.0, Component v1.9)
> MCA pml: bfo (MCA v2.0, API v2.0, Component v1.9)
> MCA pml: dr (MCA v2.0, API v2.0, Component v1.9)
> MCA pml: ob1 (MCA v2.0, API v2.0, Component v1.9)
> MCA rcache: vma (MCA v2.0, API v2.0, Component v1.9)
> MCA sharedfp: dummy (MCA v2.0, API v2.0, Component v1.9)
> MCA topo: unity (MCA v2.0, API v2.0, Component v1.9)
>
>
> I would be grateful, if somebody can fix the problems. Thank you
> very much for any help in advance.
>
>
> Kind regards
>
> Siegmar
>
> _______________________________________________
> users mailing list
> users_at_[hidden]
> http://www.open-mpi.org/mailman/listinfo.cgi/users