Not exactly. I have 16 core nodes. Even if I run all 9 ranks on the same
node it fails (with --mca btl sm,self). I also tried running on different
nodes (3 nodes, 3 ranks each on each node) with openib and tcp - the same
effect. Also as I wrote in another message I could see this effect on vbox
with CentOS 5.3 (1 cores on guest, 4 cores on host, no network). So possibly
this is something OS specific? Will try on Ubuntu and share the results.
> -----Original Message-----
> From: users-bounces_at_[hidden] [mailto:users-bounces_at_[hidden]] On
> Behalf Of Peter Kjellstrom
> Sent: Wednesday, May 25, 2011 9:03 PM
> To: users_at_[hidden]
> Subject: Re: [OMPI users] MPI_Allgather with derived type crash
> Would 8 happen to be the number of cores you have per node so what
> we're seeing is: single node OK, multi node FAIL?
> If so what kind of inter node network are you (trying to) use(ing)?