Some additional data:
Without threads it still hangs, similar behavior as before.
All of the tests were run on a system running FC11 with X5550 processors.
I just reran on a node of a RHEL 5.3 cluster with E5530 processors (dual Nehalam):
- openmpi 1.3.4 and gcc 4.1.2
- No issues: connectivity_c works through np = 128
- openmpi 1.3.4 and gcc 4.4.0
- Hangs as before
Anything else you want me to try;-)?
On Dec 10, 2009, at 5:01 PM, Gus Correa wrote:Note, too, that --enable-mpi-threads "works" but I would not say that it is production-quality hardened yet. IBM is looking into thread safety issues to harden up this code. If the same hangs can be observed without --enable-mpi-threads, that would be a good data point.
> > Just a quick interjection, I also have a dual-quad Nehalem system, HT
> > on, 24GB ram, hand compiled 1.3.4 with options: --enable-mpi-threads
> > --enable-mpi-f77=no --with-openib=no
> > With v1.3.4 I see roughly the same behavior, hello, ring work,
> > connectivity fails randomly with np >= 8. Turning on -v increased the
> > success, but still hangs. np = 16 fails more often, and the hang is
> > random in which pair of processes are communicating.
> > However, it seems to be related to the shared memory layer problem.
> > Running with -mca btl ^sm works consistently through np = 128.
users mailing list