Open MPI logo

Open MPI User's Mailing List Archives

  |   Home   |   Support   |   FAQ   |   all Open MPI User's mailing list

Subject: Re: [OMPI users] MPI_IN_PLACE in a call to MPI_Allreduce in Fortran
From: Hugo Gagnon (opensource.openmpi_at_[hidden])
Date: 2013-09-07 10:16:35


Nope, no luck. My environment is:

OpenMPI 1.6.5
gcc 4.8.1
Mac OS 10.8

I found a ticket reporting a similar problem on OS X:

https://svn.open-mpi.org/trac/ompi/ticket/1982

It said to make sure $prefix/share/ompi/mpif90-wrapper-data.txt had the
following line:

compiler_flags=-Wl,-commons,use_dylibs

I checked mine and it does (I even tried to include it explicitly on the
command line but without success), what should I do next?

-- 
  Hugo Gagnon
On Sat, Sep 7, 2013, at 0:39, Tom Rosmond wrote:
> Just as an experiment, try replacing
> 
> use mpi
> 
>   with
> 
> include 'mpif.h'
> 
> If that fixes the problem, you can confront the  OpenMPI experts
> 
> T. Rosmond
> 
> 
> 
> On Fri, 2013-09-06 at 23:14 -0400, Hugo Gagnon wrote:
> > Thanks for the input but it still doesn't work for me...  Here's the
> > version without MPI_IN_PLACE that does work:
> > 
> > program test
> > use mpi
> > integer :: ierr, myrank, a(2), a_loc(2) = 0
> > call MPI_Init(ierr)
> > call MPI_Comm_rank(MPI_COMM_WORLD,myrank,ierr)
> > if (myrank == 0) then
> >   a_loc(1) = 1
> >   a_loc(2) = 2
> > else
> >   a_loc(1) = 3
> >   a_loc(2) = 4
> > endif
> > call MPI_Allreduce(a_loc,a,2,MPI_INTEGER,MPI_SUM,MPI_COMM_WORLD,ierr)
> > write(*,*) myrank, a(:)
> > call MPI_Finalize(ierr)
> > end program test
> > 
> > $ openmpif90 test.f90
> > $ openmpirun -np 2 a.out
> >            0           4           6
> >            1           4           6
> > 
> > Now I'd be curious to know why your OpenMPI implementation handles
> > MPI_IN_PLACE correctly and not mine!
> > 
> 
> 
> _______________________________________________
> users mailing list
> users_at_[hidden]
> http://www.open-mpi.org/mailman/listinfo.cgi/users