Open MPI logo

Open MPI User's Mailing List Archives

  |   Home   |   Support   |   FAQ   |   all Open MPI User's mailing list

Subject: Re: [OMPI users] mpirun jobs only one single node
From: Ralph Castain (rhc_at_[hidden])
Date: 2010-06-15 09:49:38


That's what I suspected. I suggest you talk to your sys admin about how PBS is configured - looks like you are only getting one node allocated despite your request for two. Probably something in the config needs adjusting.

On Jun 15, 2010, at 7:20 AM, Govind Songara wrote:

> I added the $PBS_NODEFILE in the script in my last email below.
> It show only one node here is the output
> ===============
> node47.beowulf.cluster node47.beowulf.cluster node47.beowulf.cluster node47.beowulf.cluster
> This job has allocated 4 nodes
> Hello World! from process 1 out of 4 on node47.beowulf.cluster
> Hello World! from process 2 out of 4 on node47.beowulf.cluster
> Hello World! from process 3 out of 4 on node47.beowulf.cluster
> Hello World! from process 0 out of 4 on node47.beowulf.cluster
> ===============
>
> On 15 June 2010 13:41, Ralph Castain <rhc_at_[hidden]> wrote:
> Look at the contents of $PBS_NODEFILE and see how many nodes it contains.
>
> On Jun 15, 2010, at 3:56 AM, Govind Songara wrote:
>
>> Hi,
>>
>> I have using openmpi build with tm support
>> When i run the job requesting for two nodes it run only on single node.
>> Here is my script.
>> >cat mpipbs-script.sh
>> #PBS -N mpipbs-script
>> #PBS -q short
>> ### Number of nodes: resources per node
>> ### (4 cores/node, so ppn=4 is ALL resources on the node)
>> #PBS -l nodes=2:ppn=4
>
>> echo `cat $PBS_NODEFILE`
>> NPROCS=`wc -l < $PBS_NODEFILE`
>> echo This job has allocated $NPROCS nodes
>
> /opt/openmpi-1.4.2/bin/mpirun /scratch0/gsongara/mpitest/hello
>
>
> torque config
> set queue short resources_max.nodes = 4
> set queue short resources_default.nodes = 1
> set server resources_default.neednodes = 1
> set server resources_default.nodect = 1
> set server resources_default.nodes = 1
>
> Can someone please advise if i missing anything here.
>
> Regards
> Govind
>> _______________________________________________
>> users mailing list
>> users_at_[hidden]
>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>
>
> _______________________________________________
> users mailing list
> users_at_[hidden]
> http://www.open-mpi.org/mailman/listinfo.cgi/users
>
> _______________________________________________
> users mailing list
> users_at_[hidden]
> http://www.open-mpi.org/mailman/listinfo.cgi/users