Open MPI logo

Open MPI User's Mailing List Archives

  |   Home   |   Support   |   FAQ   |   all Open MPI User's mailing list

Subject: Re: [OMPI users] What flags for configure for a single machineinstallation ?
From: Jeff Squyres (jsquyres_at_[hidden])
Date: 2009-06-04 21:23:28

On Jun 4, 2009, at 12:01 PM, DEVEL Michel wrote:

> 1°) In fact I just want to install openmpi on my machine (single i7
> 920)
> to be able to develop parallel codes (using eclipse/photran/PTP)
> that I
> will execute on a cluster later (using SGE batch queue system).
> I therefore wonder what kind of configure flags I could put to have a
> basic single-machine installation ?

Nope, you shouldn't need anything special.

> 2°) For GCC, "./configure --prefix=/usr/local --with-sge
> --enable-static" worked but when I try to statically link a test
> code by
> gfortran -m64 -O3 -fPIC -fopenmp -fbounds-check -pthread --static
> testmpirun.f -o bin/testmpirun_gfortran_static -I/usr/local/include
> -L/usr/local/lib -lmpi_f90 -lmpi_f77 -lmpi -lopen-rte -lopen-pal -ldl
> -lnsl -lutil -lm -ldl
> It fails because the link step does not find Infiniband routines
> (ibv_*).

Per the other thread, static linking with OpenFabrics is not for the
meek. See the OMPI FAQ in the OpenFabrics section for a question on
exactly this issue.

> If I use dynamical link, it works but asks me for a password when I
> try
> to do
> "/usr/bin/local/mpirun -np 4 bin/testmpirun_gfortran_static" though I
> have an a priori valid .rhosts file...

Also per the other thread, this is not a static linking/dynamic
linking issue.

> 3°) for the intel compiler suite case
> "./configure --prefix=/opt/intel/Compiler/11.0/074 --with-sge
> --enable-static CC='icc' CFLAGS=' -xHOST -ip -O3 -C' LDFLAGS='-xHOST
> -ip
> -O3 -C -static-intel' AR='ar' F77='ifort' FC='ifort' FFLAGS=' -xHOST
> -ip
> -O3 -C' FCFLAGS=' -xHOST -ip -O3 -C' CXX='icpc' CXXFLAGS=' -xHOST -ip
> -O3 -C'"
> worked but I have the same problem with missing ibv_ * routines if I
> try
> a static link
> "ifort -Bdynamic -fast -C -openmp -check noarg_temp_created
> testmpirun.f -o bin/testmpirun_ifort_dynamic
> -I/opt/intel/Compiler/11.0/074/include
> -L/opt/intel/Compiler/11.0/074/lib -lmpi_f90 -lmpi_f77 -lmpi -lopen-
> rte
> -lopen-pal -ldl -lnsl -lutil -lm -ldl"
> (Remark: If I add "-static" to LDFLAGS in configure, it stops during
> the making of opal_wrapper).

Is there a reason you need static linking? It should be tremendously
simpler to get dynamic linking working.

> If I use dynamic link, I get the executable but then
> /opt/intel/Compiler/11.0/074/bin/mpirun -np 4
> ../../bin/testmpirun_ifort_dynamic
> gives
> --------------------------------------------------------------------------
> mpirun noticed that process rank 0 with PID 16664 on node mn2s-devel
> exited on signal 11 (Segmentation fault).
> --------------------------------------------------------------------------
> 2 total processes killed (some possibly by mpirun during cleanup)

What is your MPI application? Are you able to run simple MPI
applications, such as "hello world" and "ring"? (these are in the
examples/ directory in the OMPI tarball)

Jeff Squyres
Cisco Systems