Open MPI logo

Open MPI User's Mailing List Archives

  |   Home   |   Support   |   FAQ   |   all Open MPI User's mailing list

From: Ralph H Castain (rhc_at_[hidden])
Date: 2007-01-17 09:00:29

Hi Eddie

Open MPI needs to create a temporary file system ­ what we call our ³session
directory² - where it stores things like the shared memory file. From this
output, it appears that your /tmp directory is ³locked² to root access only.

You have three options for resolving this problem:

(a) you could make /tmp accessible to general users;

(b) you could use the ‹tmpdir xxx command line option to point Open MPI at
another directory that is accessible to the user (for example, you could use
a ³tmp² directory under the user¹s home directory); or

(c) you could set an MCA parameter OMPI_MCA_tmpdir_base to identify a
directory we can use instead of /tmp.

 If you select options (b) or (c), the only requirement is that this
location must be accessible on every node being used. Let me be clear on
this: the tmp directory must not be NSF mounted and therefore shared across
all nodes. However, each node must be able to access a location of the given
name ­ that location should be strictly local to each node.

Hope that helps

On 1/17/07 12:25 AM, "eddie168" <eddie168+ompi_user_at_[hidden]> wrote:

> Dear all,
> I have recently installed OpenMPI 1.1.2 on a OpenSSI cluster running Fedora
> core 3. I tested a simple hello world mpi program (attached) and it runs ok as
> root. However, if I run the same program under normal user, it gives the
> following error:
> [eddie_at_oceanus:~/home2/mpi_tut]$ mpirun -np 2 tut01
> [oceanus:125089] mca_common_sm_mmap_init: ftruncate failed with errno=13
> [oceanus:125089] mca_mpool_sm_init: unable to create shared memory mapping (
> /tmp/openmpi-sessions-eddie_at_localhost_0/default-universe/1/shared_mem_pool.loc
> alhost)
> --------------------------------------------------------------------------
> It looks like MPI_INIT failed for some reason; your parallel process is
> likely to abort. There are many reasons that a parallel process can
> fail during MPI_INIT; some of which are due to configuration or environment
> problems. This failure appears to be an internal failure; here's some
> additional information (which may only be relevant to an Open MPI
> developer):
> PML add procs failed
> --> Returned "Out of resource" (-2) instead of "Success" (0)
> --------------------------------------------------------------------------
> *** An error occurred in MPI_Init
> *** before MPI was initialized
> *** MPI_ERRORS_ARE_FATAL (goodbye)
> [eddie_at_oceanus:~/home2/mpi_tut]$
> Am I need to give certain permission to the user in order to oversubscribe
> processes?
> Thanks in advance,
> Eddie.
> _______________________________________________
> users mailing list
> users_at_[hidden]