You can confirm that the slowdown happen during the MPI initialization stages by profiling the application (especially the MPI_Init call).
Another possible cause of slowdown might be the communication thread in the ORTE. If it remains active outside the initialization it will definitively disturb the application, by taking away critical resources.
On Sep 4, 2013, at 05:59 , Christopher Samuel <samuel_at_[hidden]> wrote:
> -----BEGIN PGP SIGNED MESSAGE-----
> Hash: SHA1
> On 04/09/13 11:29, Ralph Castain wrote:
>> Your code is obviously doing something much more than just
>> launching and wiring up, so it is difficult to assess the
>> difference in speed between 1.6.5 and 1.7.3 - my guess is that it
>> has to do with changes in the MPI transport layer and nothing to do
>> with PMI or not.
> I'm testing with what would be our most used application in aggregate
> across our systems, the NAMD molecular dynamics code from here:
> so yes, you're quite right, it's doing a lot more than that and has a
> reputation for being a *very* chatty MPI code.
> For comparison whilst users see GROMACS also suffer with srun under
> 1.6.5 they don't see anything like the slow down that NAMD gets.
> All the best,
> - --
> Christopher Samuel Senior Systems Administrator
> VLSCI - Victorian Life Sciences Computation Initiative
> Email: samuel_at_[hidden] Phone: +61 (0)3 903 55545
> http://www.vlsci.org.au/ http://twitter.com/vlsci
> -----BEGIN PGP SIGNATURE-----
> Version: GnuPG v1.4.11 (GNU/Linux)
> Comment: Using GnuPG with Thunderbird - http://www.enigmail.net/
> -----END PGP SIGNATURE-----
> devel mailing list