This web mail archive is frozen.
This page is part of a frozen web archive of this mailing list.
You can still navigate around this archive, but know that no new mails
have been added to it since July of 2016.
Click here to be taken to the new web archives of this list; it includes all the mails that are in this frozen archive plus all new mails that have been sent to the list since it was migrated to the new archives.
I am testing the one-sided message passing (mpi_put, mpi_get) that is
now supported in the 1.1 release. It seems to work OK for some simple
test codes, but when I run my big application, it fails. This
application is a large weather model that runs operationally on the SGI
Origin 3000, using the native one-sided message passing that has been
supported on that system for many years. At least on that architecture,
the code always runs correctly for processor numbers up to 480. On the
O3K a requirement for the one-sided communication to work correctly is
to use 'mpi_win_create' to define the RMA 'windows' in symmetric
locations on all processors, i.e. the same 'place' in memory on each
processor. This can be done with static memory, i.e. , in common; or
on the 'symmetric heap', which is defined via environment variables. In
my application the latter method is used. I define several of these
'windows' on the symmetric heap, each with a unique handle.
Before I spend my time trying to diagnose this problem further, I need
as much information about the OpenMPI one-sided implementation as
available. Do you have a similar requirement or criteria for symmetric
memory for the RMA windows? Are there runtime parameters that I should
be using that are unique to one-sided message passing with OpenMPI? Any
other information will certainly be appreciated.
I have attached the logfiles and ompi output for my installation. I am
running with Fedora 4 linux on a 2-processor (each dual-core) Opteron
system, using the Intel 9.1.033 F90 compiler.