The Open MPI Team, representing a consortium of research, academic,
and industry partners, is pleased to announce the release of Open
MPI version 1.1.2. This release is mainly a bug fix release over
the the v1.1.1 release, but there are a few minor new features.
Version 1.1.2 can be downloaded from the main Open MPI web site or
any of its mirrors (mirrors will be updating shortly).
We strongly recommend that all users upgrade to version 1.1.2 if
Here are a list of changes in v1.1.2 as compared to v1.1.1:
- Really fix Fortran status handling in MPI_WAITSOME and MPI_TESTSOME.
- Various datatype fixes, reported by several users as causing
failures in the BLACS testing suite. Thanks to Harald Forbert, Ake
Sandgren and, Michael Kluskens for reporting the problem.
- Correctness and performance fixes for heterogeneous environments.
- Fixed a error in command line parsing on some platforms (causing
mpirun to crash without doing anything).
- Fix for initialization hangs on 64 bit Mac OS X PowerPC systems.
- Fixed some memory allocation problems in mpirun that could cause
random problems if "-np" was not specified on the command line.
- Add Kerberos authentication support for XGrid.
- Added LoadLeveler support for jobs larger than 128 tasks.
- Fix for large-sized Fortran LOGICAL datatypes.
- Fix various error checking in MPI_INFO_GET_NTHKEY and
MPI_GROUP_TRANSLATE_RANKS, and some collective operations
(particularly with regards to MPI_IN_PLACE). Thanks to Lisandro
Dalcin for reporting the problems.
- Fix receiving messages to buffers allocated by MPI_ALLOC_MEM.
- Fix a number of race conditions with the MPI-2 Onesided
- Fix the "tuned" collective componenete where some cases where
MPI_BCAST could hang.
- Update TCP support to support non-uniform TCP environments.
- Allow the "poe" RAS component to be built on AIX or Linux.
- Only install mpif.h if the rest of the Fortran bindings are
- Fixes for BProc node selection.
- Add some missing Fortran MPI-2 IO constants.
Server Virtualization Business Unit