Open MPI logo

Open MPI User's Mailing List Archives

  |   Home   |   Support   |   FAQ   |   all Open MPI User's mailing list

Subject: Re: [OMPI users] mpirun example program fail on multiple nodes- unable to launch specified application on client node
From: qing.pang (qing.pang_at_[hidden])
Date: 2009-11-06 11:44:16

This sounds like what I need exactly. But when I tried it, I got the
following problem - I'm working on my Desktop, and trying to preload
the executable to my laptop.

gordon_at_gordon-desktop:~/Desktop/openmpi-1.3.3/examples$ mpirun
-machinefile machine.linux -np 2 --preload-binary $(pwd)/hello_c.out
gordon_at_gordon-desktop's password:
WARNING: Remote peer ([[18118,0],1]) failed to preload a file.

Exit Status: 256
Local  File: /tmp/openmpi-sessions-gordon_at_gordon-laptop_0/18118/0/hello_c.out
Remote File: /home/gordon/Desktop/openmpi-1.3.3/examples/hello_c.out
  scp  gordon-desktop:/home/gordon/Desktop/openmpi-1.3.3/examples/hello_c.out

Will continue attempting to launch the process(es).
mpirun was unable to launch the specified application as it could not access
or execute an executable:

Executable: /home/gordon/Desktop/openmpi-1.3.3/examples/hello_c.out
Node: node1

while attempting to start process rank 1.

I typed in my password for master node account when asked for it. But
why I was asked for my password on master node - I am working under
this account anyway?


On Fri, Nov 6, 2009 at 7:09 AM, Josh Hursey <jjhursey_at_[hidden]> wrote:
> As an alternative technique for distributing the binary, you could ask Open MPI's runtime to do it for you (made available in the v1.3 series). You still need to make sure that the same version of Open is installed on all nodes, but if you pass the --preload-binary option to mpirun the runtime environment will distribute the binary across the machine (staging it to a temporary directory) before launching it.
> You can do the same with any arbitrary set of files or directories (comma separated) using the --preload-files option as well.
> If you type 'mpirun --help' the options that you are looking for are:
> --------------------
>   --preload-files <arg0>
>   --preload-files-dest-dir <arg0>
>                         with --preload-files. By default the absolute and
>                         relative paths provided by --preload-files are
> -s|--preload-binary      Preload the binary on the remote machine before
> --------------------
> -- Josh
> On Nov 5, 2009, at 6:56 PM, Terry Frankcombe wrote:
>> For small ad hoc COWs I'd vote for sshfs too.  It may well be as slow as
>> a dog, but it actually has some security, unlike NFS, and is a doddle to
>> make work with no superuser access on the server, unlike NFS.
>> On Thu, 2009-11-05 at 17:53 -0500, Jeff Squyres wrote:
>>> On Nov 5, 2009, at 5:34 PM, Douglas Guptill wrote:
>>>> I am currently using sshfs to mount both OpenMPI and my application on
>>>> the "other" computers/nodes.  The advantage to this is that I have
>>>> only one copy of OpenMPI and my application.  There may be a
>>>> performance penalty, but I haven't seen it yet.
>>> For a small number of nodes (where small <=32 or sometimes even <=64),
>>> I find that simple NFS works just fine.  If your apps aren't IO
>>> intensive, that can greatly simplify installation and deployment of
>>> both Open MPI and your MPI applications IMNSHO.
>>> But -- every app is different.  :-)  YMMV.
>> _______________________________________________
>> users mailing list
>> users_at_[hidden]
> _______________________________________________
> users mailing list
> users_at_[hidden]