Open MPI logo

Open MPI User's Mailing List Archives

  |   Home   |   Support   |   FAQ   |   all Open MPI User's mailing list

Subject: Re: [OMPI users] open mpi 1.6 with intel compilers
From: Reuti (reuti_at_[hidden])
Date: 2012-10-29 10:54:21


Am 29.10.2012 um 14:49 schrieb Giuseppe P.:

> Thank you very much guys. Now a more serious issue:
>
> I am using mpi with laamps (a molecular dynamics package) on a single Rack Dell server Poweredge R810
> (4 eight-core processors, 128 Gb RAM memory).
> I am now potentially interested into buying the Intel MPI 4.1 libraries, and I am trying them by exploiting the
> 30 days trial. However, I am not experiencing any significant improved performance by using
> the Intel MPI libraries with respect to the Open Mpi (compiled with the Itel compilers).
>
> Here there is the (makefile) working configuration for the Intel MPI 4.1 compilers:
> CC = /opt/intel/impi/4.1.0.024/intel64/bin/mpiicpc
> CCFLAGS = -O -DMPICH_IGNORE_CXX_SEEK -DMPICH_SKIP_MPICXX
>
> And here there is the Open Mpi one:
> CC = /usr/local/bin/mpicc
> CCFLAGS = -O -mpicc
>
> I also tried the flag -O3 but I detected no significant differences in performance.
> Now, I would be considering the Intel Mpi libraries, provided this would bring to a
> significant increase in performance with respect to Open Mpi.

Why - because of -O3 (which I would consider dangerous) or because it's from Intel? Intel MPI is based on MPICH2, not Open MPI.

> I have evidence that there is room to improve because laamps under the same conditions
> and on an HP Z650 with two 6-core processors (the clock frequency is the same though,
> and tests for comparison were done on parallel runs using 8 cores), improves of nearly the 70%
> by using the proprietary HP MPI libraries.

NB: They made their way to Platform Computing and then to IBM.

-- Reuti

> Kind regards
>
> Giuseppe
>
>
> 2012/10/27 Ralph Castain <rhc_at_[hidden]>
> The reason is that you aren't actually running Open MPI - those error messages are coming from MPICH. Check your path and ensure you put the OMPI install location first, or use the absolute path to the OMPI mpirun
>
> On Oct 27, 2012, at 8:46 AM, Giuseppe P. <istruzione_at_[hidden]> wrote:
>
>> Hello!
>>
>> I have built open mpi 1.6 with Intel compilers (2013 versions). Compilation was smooth, however even when I try to execute
>> the simple program hello.c:
>>
>> mpirun -np 4 ./hello_c.x
>> [mpiexec_at_claudio.ukzn] HYDU_create_process (./utils/launch/launch.c:102): execvp error on file /opt/intel/composer_xe_2013.0.079/mpirt/bin/intel64/pmi_proxy (No such file or directory)
>> [mpiexec_at_claudio.ukzn] HYD_pmcd_pmiserv_proxy_init_cb (./pm/pmiserv/pmiserv_cb.c:1177): assert (!closed) failed
>> [mpiexec_at_claudio.ukzn] HYDT_dmxu_poll_wait_for_event (./tools/demux/demux_poll.c:77): callback returned error status
>> [mpiexec_at_claudio.ukzn] HYD_pmci_wait_for_completion (./pm/pmiserv/pmiserv_pmci.c:358): error waiting for event
>> [mpiexec_at_claudio.ukzn] main (./ui/mpich/mpiexec.c:689): process manager error waiting for completion
>>
>> Before that, there was an additional error, since also the file mpivars.sh was not present in /opt/intel/composer_xe_2013.0.079/mpirt/bin/intel64/.
>> Even though I managed to create one and it worked:
>>
>> #!/bin/bash
>>
>> if [ -z "`echo $PATH | grep /usr/local/bin`" ]; then
>> export PATH=/usr/local/bin:$PATH
>> fi
>>
>> if [ -z "`echo $LD_LIBRARY_PATH | grep /usr/local/lib`" ]; then
>> if [ -n "$LD_LIBRARY_PATH" ]; then
>> export LD_LIBRARY_PATH=/usr/local/lib:$LD_LIBRARY_PATH
>> else
>> export LD_LIBRARY_PATH=/usr/local/lib
>> fi
>> fi
>>
>> I do not have any clue about how to generate the file pmi_proxy.
>>
>> Thank you in advance for your help!
>>
>> _______________________________________________
>>
>> users mailing list
>> users_at_[hidden]
>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>
>
> _______________________________________________
> users mailing list
> users_at_[hidden]
> http://www.open-mpi.org/mailman/listinfo.cgi/users
>
> _______________________________________________
> users mailing list
> users_at_[hidden]
> http://www.open-mpi.org/mailman/listinfo.cgi/users