Open MPI logo

Open MPI User's Mailing List Archives

  |   Home   |   Support   |   FAQ   |   all Open MPI User's mailing list

Subject: Re: [OMPI users] MPI_Allreduce on local machine
From: Jeff Squyres (jsquyres_at_[hidden])
Date: 2010-07-28 08:57:21


I don't have the intel compilers on my Mac, but I'm unable to replicate this issue on Linux with the intel compilers v11.0.

Can you get a corefile to see a backtrace where it died in Open MPI's allreduce?

How exactly did you configure your Open MPI, and how exactly did you compile / run your sample application?

On Jul 27, 2010, at 10:35 PM, Hugo Gagnon wrote:

> I did and it runs now, but the result is wrong: outside is still 1.d0,
> 2.d0, 3.d0, 4.d0, 5.d0
> How can I make sure to compile OpenMPI so that datatypes such as
> mpi_double_precision behave as they "should"?
> Are there flags during the OpenMPI building process or something?
> Thanks,
> --
> Hugo Gagnon
>
>
> On Tue, 27 Jul 2010 09:06 -0700, "David Zhang" <solarbikedz_at_[hidden]>
> wrote:
> > Try mpi_real8 for the type in allreduce
> >
> > On 7/26/10, Hugo Gagnon <sourceforge.openmpi_at_[hidden]> wrote:
> > > Hello,
> > >
> > > When I compile and run this code snippet:
> > >
> > > 1 program test
> > > 2
> > > 3 use mpi
> > > 4
> > > 5 implicit none
> > > 6
> > > 7 integer :: ierr, nproc, myrank
> > > 8 integer, parameter :: dp = kind(1.d0)
> > > 9 real(kind=dp) :: inside(5), outside(5)
> > > 10
> > > 11 call mpi_init(ierr)
> > > 12 call mpi_comm_size(mpi_comm_world, nproc, ierr)
> > > 13 call mpi_comm_rank(mpi_comm_world, myrank, ierr)
> > > 14
> > > 15 inside = (/ 1, 2, 3, 4, 5 /)
> > > 16 call mpi_allreduce(inside, outside, 5, mpi_double_precision,
> > > mpi_sum, mpi_comm_world, ierr)
> > > 17
> > > 18 print*, myrank, inside
> > > 19 print*, outside
> > > 20
> > > 21 call mpi_finalize(ierr)
> > > 22
> > > 23 end program test
> > >
> > > I get the following error, with say 2 processors:
> > >
> > > forrtl: severe (174): SIGSEGV, segmentation fault occurred
> > > Image PC Routine Line
> > > Source
> > > libmpi.0.dylib 00000001001BB4B7 Unknown Unknown
> > > Unknown
> > > libmpi_f77.0.dyli 00000001000AF046 Unknown Unknown
> > > Unknown
> > > a.out 0000000100000CE2 _MAIN__ 16
> > > test.f90
> > > a.out 0000000100000BDC Unknown Unknown
> > > Unknown
> > > a.out 0000000100000B74 Unknown Unknown
> > > Unknown
> > > forrtl: severe (174): SIGSEGV, segmentation fault occurred
> > > Image PC Routine Line
> > > Source
> > > libmpi.0.dylib 00000001001BB4B7 Unknown Unknown
> > > Unknown
> > > libmpi_f77.0.dyli 00000001000AF046 Unknown Unknown
> > > Unknown
> > > a.out 0000000100000CE2 _MAIN__ 16
> > > test.f90
> > > a.out 0000000100000BDC Unknown Unknown
> > > Unknown
> > > a.out 0000000100000B74 Unknown Unknown
> > > Unknown
> > >
> > > on my iMac having compiled OpenMPI with ifort according to:
> > > http://software.intel.com/en-us/articles/performance-tools-for-software-developers-building-open-mpi-with-the-intel-compilers/
> > >
> > > Note that the above code snippet runs fine on my school parallel cluster
> > > where ifort+intelmpi is installed.
> > > Is there something special about OpenMPI's MPI_Allreduce function call
> > > that I should be aware of?
> > >
> > > Thanks,
> > > --
> > > Hugo Gagnon
> > >
> > > _______________________________________________
> > > users mailing list
> > > users_at_[hidden]
> > > http://www.open-mpi.org/mailman/listinfo.cgi/users
> > >
> >
> > --
> > Sent from my mobile device
> >
> > David Zhang
> > University of California, San Diego
> > _______________________________________________
> > users mailing list
> > users_at_[hidden]
> > http://www.open-mpi.org/mailman/listinfo.cgi/users
> >
> --
> Hugo Gagnon
>
> _______________________________________________
> users mailing list
> users_at_[hidden]
> http://www.open-mpi.org/mailman/listinfo.cgi/users
>

-- 
Jeff Squyres
jsquyres_at_[hidden]
For corporate legal information go to:
http://www.cisco.com/web/about/doing_business/legal/cri/