Open MPI logo

Open MPI User's Mailing List Archives

  |   Home   |   Support   |   FAQ   |   all Open MPI User's mailing list

From: Maestas, Christopher Daniel (cdmaest_at_[hidden])
Date: 2006-10-16 09:59:46

How fast/well are MPI collectives implemented in ompi?
I'm running the Intel MPI 1.1. benchmarks and seeing the need to set
wall clock times > 12 hours for run sizes of 200 and 300 nodes for 1ppn
and 2ppn cases. The collective tests that usually pass in 2ppn cases:
 Barrier, Reduce scatter, allreduce, bcast

The ones that take long or never run:
 Allgather, alltoall, allgatherv