Open MPI logo

Open MPI User's Mailing List Archives

  |   Home   |   Support   |   FAQ   |   all Open MPI User's mailing list

Subject: Re: [OMPI users] Problems with gridengine integration on RHEL 6
From: Dave Love (d.love_at_[hidden])
Date: 2012-02-16 06:05:54


Brian McNally <bmcnally_at_[hidden]> writes:

> Hi Dave,
>
> I looked through the INSTALL, VERSION, NEWS, and README files in the
> 1.5.4 openmpi tarball but didn't see what you were referring to.

I can't access the web site, but there's an item in the notes on the
download page about the bug. It must also be in the mail archive in a
thread with me included.

> Are
> you suggesting that I launch mpirun similar to this?
>
> mpirun -mca plm ^rshd ...?

Yes, or put it in openmpi-mca-params.conf. It's harmless with 1.4.

> What I meant by "the same parallel environment setup" was that the PE
> in SGE was defined the same way:
>
> $ qconf -sp orte
> pe_name orte
> slots 9999
> user_lists NONE
> xuser_lists NONE
> start_proc_args /bin/true
> stop_proc_args /bin/true
> allocation_rule $round_robin
> control_slaves TRUE
> job_is_first_task FALSE
> urgency_slots min
> accounting_summary FALSE
>
> Even though I have RHEL 5 and RHEL 6 nodes in the same cluster they
> never run the same MPI job; it's always either all RHEL 5 nodes or all
> RHEL 6.

OK. (I'd expect a separate PE for each node set to enforce that.)