This web mail archive is frozen.
This page is part of a frozen web archive of this mailing list.
You can still navigate around this archive, but know that no new mails
have been added to it since July of 2016.
Click here to be taken to the new web archives of this list; it includes all the mails that are in this frozen archive plus all new mails that have been sent to the list since it was migrated to the new archives.
How did you configure OMPI? If you add --display-allocation to your cmd line, does it show all the nodes?
On Jan 24, 2013, at 6:34 AM, Sabuj Pattanayek <sabujp_at_[hidden]> wrote:
> I'm submitting a job through torque/PBS, the head node also runs the
> Moab scheduler, the .pbs file has this in the resources line :
> #PBS -l nodes=2:ppn=4
> I've also tried something like :
> #PBS -l procs=56
> and at the end of script I'm running :
> mpirun -np 8 cat /dev/urandom > /dev/null
> mpirun -np 56 cat /dev/urandom > /dev/null
> ...depending on how many processors I requested. The job starts,
> $PBS_NODEFILE has the nodes that the job was assigned listed, but all
> the cat's are piled onto the first node. Any idea how I can get this
> to submit jobs across multiple nodes? Note, I have OSU mpiexec working
> without problems with mvapich and mpich2 on our cluster to launch jobs
> across multiple nodes.
> users mailing list