Open MPI logo

Open MPI User's Mailing List Archives

  |   Home   |   Support   |   FAQ   |  

This web mail archive is frozen.

This page is part of a frozen web archive of this mailing list.

You can still navigate around this archive, but know that no new mails have been added to it since July of 2016.

Click here to be taken to the new web archives of this list; it includes all the mails that are in this frozen archive plus all new mails that have been sent to the list since it was migrated to the new archives.

Subject: Re: [OMPI users] mpirun runs in serial even I set np to several processors
From: Gus Correa (gus_at_[hidden])
Date: 2014-04-14 16:09:36


Djordje

Your WRF configure file seems to use mpif90 and mpicc (line 115 &
following).
In addition, it also seems to have DISABLED OpenMP (NO TRAILING "I")
(lines 109-111, where OpenMP stuff is commented out).
So, it looks like to me your intent was to compile with MPI.

Whether it is THIS MPI (OpenMPI) or another MPI (say MPICH, or MVAPICH,
or Intel MPI, or Cray, or ...) only your environment can tell.

What do you get from these commands:

which mpirun
which mpif90
which mpicc

I never built WRF here (but other people here use it).
Which input do you provide to the command that generates the configure
script that you sent before?
Maybe the full command line will shed some light on the problem.

I hope this helps,
Gus Correa

On 04/14/2014 03:11 PM, Djordje Romanic wrote:
> to get help :)
>
>
>
> On Mon, Apr 14, 2014 at 3:11 PM, Djordje Romanic <djordje8_at_[hidden]
> <mailto:djordje8_at_[hidden]>> wrote:
>
> Yes, but I was hoping to get. :)
>
>
> On Mon, Apr 14, 2014 at 3:02 PM, Jeff Squyres (jsquyres)
> <jsquyres_at_[hidden] <mailto:jsquyres_at_[hidden]>> wrote:
>
> If you didn't use Open MPI, then this is the wrong mailing list
> for you. :-)
>
> (this is the Open MPI users' support mailing list)
>
>
> On Apr 14, 2014, at 2:58 PM, Djordje Romanic <djordje8_at_[hidden]
> <mailto:djordje8_at_[hidden]>> wrote:
>
> > I didn't use OpenMPI.
> >
> >
> > On Mon, Apr 14, 2014 at 2:37 PM, Jeff Squyres (jsquyres)
> <jsquyres_at_[hidden] <mailto:jsquyres_at_[hidden]>> wrote:
> > This can also happen when you compile your application with
> one MPI implementation (e.g., Open MPI), but then mistakenly use
> the "mpirun" (or "mpiexec") from a different MPI implementation
> (e.g., MPICH).
> >
> >
> > On Apr 14, 2014, at 2:32 PM, Djordje Romanic
> <djordje8_at_[hidden] <mailto:djordje8_at_[hidden]>> wrote:
> >
> > > I compiled it with: x86_64 Linux, gfortran compiler with
> gcc (dmpar). dmpar - distributed memory option.
> > >
> > > Attached is the self-generated configuration file. The
> architecture specification settings start at line 107. I didn't
> use Open MPI (shared memory option).
> > >
> > >
> > > On Mon, Apr 14, 2014 at 1:23 PM, Dave Goodell (dgoodell)
> <dgoodell_at_[hidden] <mailto:dgoodell_at_[hidden]>> wrote:
> > > On Apr 14, 2014, at 12:15 PM, Djordje Romanic
> <djordje8_at_[hidden] <mailto:djordje8_at_[hidden]>> wrote:
> > >
> > > > When I start wrf with mpirun -np 4 ./wrf.exe, I get this:
> > > > -------------------------------------------------
> > > > starting wrf task 0 of 1
> > > > starting wrf task 0 of 1
> > > > starting wrf task 0 of 1
> > > > starting wrf task 0 of 1
> > > > -------------------------------------------------
> > > > This indicates that it is not using 4 processors, but 1.
> > > >
> > > > Any idea what might be the problem?
> > >
> > > It could be that you compiled WRF with a different MPI
> implementation than you are using to run it (e.g., MPICH vs.
> Open MPI).
> > >
> > > -Dave
> > >
> > > _______________________________________________
> > > users mailing list
> > > users_at_[hidden] <mailto:users_at_[hidden]>
> > > http://www.open-mpi.org/mailman/listinfo.cgi/users
> > >
> > > <configure.wrf>_______________________________________________
> > > users mailing list
> > > users_at_[hidden] <mailto:users_at_[hidden]>
> > > http://www.open-mpi.org/mailman/listinfo.cgi/users
> >
> >
> > --
> > Jeff Squyres
> > jsquyres_at_[hidden] <mailto:jsquyres_at_[hidden]>
> > For corporate legal information go to:
> http://www.cisco.com/web/about/doing_business/legal/cri/
> >
> > _______________________________________________
> > users mailing list
> > users_at_[hidden] <mailto:users_at_[hidden]>
> > http://www.open-mpi.org/mailman/listinfo.cgi/users
> >
> > _______________________________________________
> > users mailing list
> > users_at_[hidden] <mailto:users_at_[hidden]>
> > http://www.open-mpi.org/mailman/listinfo.cgi/users
>
>
> --
> Jeff Squyres
> jsquyres_at_[hidden] <mailto:jsquyres_at_[hidden]>
> For corporate legal information go to:
> http://www.cisco.com/web/about/doing_business/legal/cri/
>
> _______________________________________________
> users mailing list
> users_at_[hidden] <mailto:users_at_[hidden]>
> http://www.open-mpi.org/mailman/listinfo.cgi/users
>
>
>
>
>
> _______________________________________________
> users mailing list
> users_at_[hidden]
> http://www.open-mpi.org/mailman/listinfo.cgi/users
>