Open MPI logo

Hardware Locality Development Mailing List Archives

  |   Home   |   Support   |   FAQ   |   all Hardware Locality Development mailing list

Subject: Re: [hwloc-devel] Fwd: BGQ empty topology with MPI
From: Christopher Samuel (samuel_at_[hidden])
Date: 2012-03-25 23:16:05

Hash: SHA1

On 25/03/12 09:04, Daniel Ibanez wrote:

> Additional printfs confirm that with MPI in the code,
> hwloc_accessat succeeds on the various /sys/ directories, but the
> overall procedure for getting PUs from these fails. Without MPI,
> access to /sys/ directories fails but the fallback
> hwloc_setup_pu_level works.

Sounds like your I/O with MPI is getting redirected to the I/O node
(and hence finding /sys from the Linux kernel there) but when you're
running without MPI it's trying to open files on the compute node and
the CNK isn't presenting the /sys directories, causing it to fall back.

I've run lstopo on our BG/P and I get to see the 4 cores there whether
it's the stock code or if I add an MPI_Init() to the start. The
output from lstopo when built with --enable-debug confirms it's
reporting kernel and hostname info from the I/O node associated with
the block:

Machine#0(Backend=Linux OSName=CNK OSRelease= OSVersion=1 Architecture=BGP) [...]

It might be interesting to build something like ls with the BG/Q
compilers to see if you can run it on a compute node to see what /proc
or /sys look like in each case.

- --
    Christopher Samuel - Senior Systems Administrator
 VLSCI - Victorian Life Sciences Computation Initiative
 Email: samuel_at_[hidden] Phone: +61 (0)3 903 55545

Version: GnuPG v1.4.11 (GNU/Linux)
Comment: Using GnuPG with Mozilla -