Open MPI logo

Open MPI User's Mailing List Archives

  |   Home   |   Support   |   FAQ   |   all Open MPI User's mailing list

Subject: [OMPI users] Openmpi and processor affinity
From: Iftikhar Rathore (irathore_at_[hidden])
Date: 2009-06-02 19:30:59


Hi
We are using openmpi version 1.2.8 (packaged with ofed-1.4). I am trying
to run hpl-2.0 (linpak). We have two intel quad core CPU's in all our
server (8 total cores) and all hosts in the hostfile have lines that
look like "10.100.0.227 slots=8max_slots=8".

Now when I use mpirun (even with --mca mpi_paffinity_alone 1) it does
not keep the affinity, the processes seem to gravitate towards first
four cores (using top and hitting 1). I know I do have MCA paffinity
available.

[root_at_devi DLR_WB_88]# ompi_info | grep paffinity
[devi.cisco.com:26178] mca: base: component_find: unable to open btl openib: file not found (ignored)
           MCA paffinity: linux (MCA v1.0, API v1.0, Component v1.2.8)

The command line I am using is:

# mpirun -nolocal -np 896 -v --mca mpi_paffinity_alone 1 -hostfile /mnt/apps/hosts/896_8slots /mnt/apps/bin/xhpl

Am I doing something wrong and is there a way to confirm cpu affinity besides hitting "1" on top.

[root_at_devi DLR_WB_88]# mpirun -V
mpirun (Open MPI) 1.2.8

Report bugs to http://www.open-mpi.org/community/help/

-- 
Iftikhar Rathore
Technical Marketing Engineer
Server Access Virtualization BU.
Cisco Systems, Inc.
Phone:  +1 408 853 5322
Mobile: +1 636 675 2982