Really have to thank you for your persistence, Paul. Truly appreciated.

Glad to hear you can run ring_c. I'm going to let Jeff handle the path.c and ANON issue as those are both configury things he was working today.

Not sure why the shmem fortran examples would try to build - will pass that off to Jeff as well (sorry Jeff!)

On Jan 9, 2014, at 9:01 PM, Paul Hargrove <phhargrove@lbl.gov> wrote:

Ralph,

My NetBSD builds fail on the new opal/util/path.c, but by adding
  #undef HAVE_STATFS
near the top of path.c I can get past that.

Next I hit the MAP_ANON-vs-MAP_ANONYMOUS issue and fixed that manually.

Next I hit the attempt to build shmem fortran examples despite the lack of a fortran compiler.

I *was* able to finally confirm that I can now run ring_c.

-Paul


On Thu, Jan 9, 2014 at 12:07 PM, Paul Hargrove <phhargrove@lbl.gov> wrote:
Ralph,

Thanks for fielding all these issues I've been finding.
I will plan to run tonight's trunk tarball through all of the systems where I've seen any issues.

-Paul


On Thu, Jan 9, 2014 at 8:40 AM, Ralph Castain <rhc@open-mpi.org> wrote:
Should now be fixed in trunk (silently fall back to not binding if cores not found) - scheduled for 1.7.4. If you could test the next trunk tarball, that would help as I can't actually test it on my machines


On Jan 9, 2014, at 6:25 AM, Ralph Castain <rhc@open-mpi.org> wrote:

I see the issue - there are no "cores" on this topology, only "pu's", so "bind-to core" is going to fail even though binding is supported. Will adjust.

Thanks!

On Jan 8, 2014, at 9:06 PM, Paul Hargrove <phhargrove@lbl.gov> wrote:

Requested verbose output below.
-Paul

-bash-4.2$ mpirun -mca ess_base_verbose 10 -np 1 examples/ring_c
[pcp-j-17:02150] mca: base: components_register: registering ess components
[pcp-j-17:02150] mca: base: components_register: found loaded component env
[pcp-j-17:02150] mca: base: components_register: component env has no register or open function
[pcp-j-17:02150] mca: base: components_register: found loaded component hnp
[pcp-j-17:02150] mca: base: components_register: component hnp has no register or open function
[pcp-j-17:02150] mca: base: components_register: found loaded component singleton
[pcp-j-17:02150] mca: base: components_register: component singleton register function successful
[pcp-j-17:02150] mca: base: components_register: found loaded component tool
[pcp-j-17:02150] mca: base: components_register: component tool has no register or open function
[pcp-j-17:02150] mca: base: components_open: opening ess components
[pcp-j-17:02150] mca: base: components_open: found loaded component env
[pcp-j-17:02150] mca: base: components_open: component env open function successful
[pcp-j-17:02150] mca: base: components_open: found loaded component hnp
[pcp-j-17:02150] mca: base: components_open: component hnp open function successful
[pcp-j-17:02150] mca: base: components_open: found loaded component singleton
[pcp-j-17:02150] mca: base: components_open: component singleton open function successful
[pcp-j-17:02150] mca: base: components_open: found loaded component tool
[pcp-j-17:02150] mca: base: components_open: component tool open function successful
[pcp-j-17:02150] mca:base:select: Auto-selecting ess components
[pcp-j-17:02150] mca:base:select:(  ess) Querying component [env]
[pcp-j-17:02150] mca:base:select:(  ess) Skipping component [env]. Query failed to return a module
[pcp-j-17:02150] mca:base:select:(  ess) Querying component [hnp]
[pcp-j-17:02150] mca:base:select:(  ess) Query of component [hnp] set priority to 100
[pcp-j-17:02150] mca:base:select:(  ess) Querying component [singleton]
[pcp-j-17:02150] mca:base:select:(  ess) Skipping component [singleton]. Query failed to return a module
[pcp-j-17:02150] mca:base:select:(  ess) Querying component [tool]
[pcp-j-17:02150] mca:base:select:(  ess) Skipping component [tool]. Query failed to return a module
[pcp-j-17:02150] mca:base:select:(  ess) Selected component [hnp]
[pcp-j-17:02150] mca: base: close: component env closed
[pcp-j-17:02150] mca: base: close: unloading component env
[pcp-j-17:02150] mca: base: close: component singleton closed
[pcp-j-17:02150] mca: base: close: unloading component singleton
[pcp-j-17:02150] mca: base: close: component tool closed
[pcp-j-17:02150] mca: base: close: unloading component tool
[pcp-j-17:02150] [[INVALID],INVALID] Topology Info:
[pcp-j-17:02150] Type: Machine Number of child objects: 2
        Name=NULL
        Backend=NetBSD
        OSName=NetBSD
        OSRelease=6.1
        OSVersion="NetBSD 6.1 (CUSTOM) #0: Fri Sep 20 13:19:58 PDT 2013 phargrov@pcp-j-17:/home/phargrov/CUSTOM"
        Architecture=i386
        Backend=x86
        Cpuset:  0x00000003
        Online:  0x00000003
        Allowed: 0x00000003
        Bind CPU proc:   TRUE
        Bind CPU thread: TRUE
        Bind MEM proc:   FALSE
        Bind MEM thread: FALSE
        Type: PU Number of child objects: 0
                Name=NULL
                Cpuset:  0x00000001
                Online:  0x00000001
                Allowed: 0x00000001
        Type: PU Number of child objects: 0
                Name=NULL
                Cpuset:  0x00000002
                Online:  0x00000002
                Allowed: 0x00000002
--------------------------------------------------------------------------
While computing bindings, we found no available cpus on
the following node:

  Node:  pcp-j-17

Please check your allocation.
--------------------------------------------------------------------------
[pcp-j-17:02150] mca: base: close: component hnp closed
[pcp-j-17:02150] mca: base: close: unloading component hnp



On Wed, Jan 8, 2014 at 8:50 PM, Ralph Castain <rhc@open-mpi.org> wrote:
Hmmm...looks to me like the code should protect against this - unless the system isn't correctly reporting binding support. Could you run this with "-mca ess_base_verbose 10"? This will output the topology we found, including the binding support (which isn't in the usual output).

On Jan 8, 2014, at 8:14 PM, Ralph Castain <rhc@open-mpi.org> wrote:

Hmmm...I see the problem. Looks like binding isn't supported on that system for some reason, so we need to turn "off" our auto-binding when we hit that condition. I'll check to see why that isn't happening (was supposed to do so)


On Jan 8, 2014, at 3:43 PM, Paul Hargrove <phhargrove@lbl.gov> wrote:

While I have yet to get a working build on NetBSD for x86-64 h/w, I *have* successfully built Open MPI's current 1.7.4rc tarball on NetBSD-6 for x86.  However, I can't *run* anything:

Attempting the ring_c example on 2 cores:
-bash-4.2$ mpirun -mca btl sm,self -np 2 examples/ring_c
--------------------------------------------------------------------------
While computing bindings, we found no available cpus on
the following node:

  Node:  pcp-j-17

Please check your allocation.
--------------------------------------------------------------------------

The failure is the same w/o "-mca btl sm,self"
Singleton runs fail just as the np=2 run did.

I've attached compressed output from "ompi_info --all".

Since this is probably an hwloc-related issue, I also build hwloc-1.7.2 from pristine sources.
I have attached compressed output of lstopo which NOTABLY indicates a failure to bind to both of the CPUs.

For now, an explicit "--bind-to none" is working for me.
Please let me know what additional info may be required.

-Paul

--
Paul H. Hargrove                          PHHargrove@lbl.gov
Future Technologies Group
Computer and Data Sciences Department     Tel: +1-510-495-2352
Lawrence Berkeley National Laboratory     Fax: +1-510-486-6900
<ompi_info-netbsd-x86.txt.bz2><lstopo172-netbsd-x86.txt.bz2>_______________________________________________
devel mailing list
devel@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/devel



_______________________________________________
devel mailing list
devel@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/devel



--
Paul H. Hargrove                          PHHargrove@lbl.gov
Future Technologies Group
Computer and Data Sciences Department     Tel: +1-510-495-2352
Lawrence Berkeley National Laboratory     Fax: +1-510-486-6900
_______________________________________________
devel mailing list
devel@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/devel



_______________________________________________
devel mailing list
devel@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/devel



--
Paul H. Hargrove                          PHHargrove@lbl.gov
Future Technologies Group
Computer and Data Sciences Department     Tel: +1-510-495-2352
Lawrence Berkeley National Laboratory     Fax: +1-510-486-6900



--
Paul H. Hargrove                          PHHargrove@lbl.gov
Future Technologies Group
Computer and Data Sciences Department     Tel: +1-510-495-2352
Lawrence Berkeley National Laboratory     Fax: +1-510-486-6900
_______________________________________________
devel mailing list
devel@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/devel