This web mail archive is frozen.
This page is part of a frozen web archive of this mailing list.
You can still navigate around this archive, but know that no new mails
have been added to it since July of 2016.
Click here to be taken to the new web archives of this list; it includes all the mails that are in this frozen archive plus all new mails that have been sent to the list since it was migrated to the new archives.
I found that MPT uses a *lot* of vmem for buffering/mem mapping. We
schedule based on requested vmem, so this can be a problem. Do you know
how vmem usage for buffering compares with OpenMPI?
From: users-bounces_at_[hidden] [mailto:users-bounces_at_[hidden]] On
Behalf Of Brian W. Barrett
Sent: Wednesday, 17 January 2007 1:49 PM
To: Open MPI Users
Subject: Re: [OMPI users] openmpi on altix
On Jan 16, 2007, at 4:29 PM, Brock Palen wrote:
> What is the state of openMPI on a sgi altix? How does it compare to
> mpt. I assume for all operations OMPI will use the sm btl thus all
> others (other than self) could be disabled. Is there any other
> tweaks users use? Or is OMPI even recommend on at Altix?
We've run Open MPI on the Altix here at Los Alamos. For point-to-
point communication, we're slightly slower than MPT. But for
collectives, we're much slower. We just haven't done any work on
shared memory collectives, especially on platforms where the memory
hierarchies are as deep as they are on the Altix. That being said,
it should work and is a viable option if there's a feature of Open
MPI that a user needs that is not available in MPT.
Open MPI Team, CCS-1
Los Alamos National Laboratory
users mailing list