Open MPI logo

Open MPI Development Mailing List Archives

  |   Home   |   Support   |   FAQ   |   all Development mailing list

From: Sridhar Chirravuri (sridhar_at_[hidden])
Date: 2005-08-17 07:23:20


Can someone reply to my mail please?

-----Original Message-----
From: devel-bounces_at_[hidden] [mailto:devel-bounces_at_[hidden]] On
Behalf Of Sridhar Chirravuri
Sent: Wednesday, August 17, 2005 4:18 PM
To: Open MPI Developers
Subject: Re: [O-MPI devel] Fwd: Regarding MVAPI Component in Open MPI

Hi,

I checked out the latest code drop r6911 today morning and ran Pallas
with in the same node (2 procs). It ran fine. I didn't see any hangs
this time whereas I could see the following statements in the pallas
output and I feel they are just warnings, which can be ignored. Am I
correct?

Request for 0 bytes (coll_basic_reduce_scatter.c, 80)
Request for 0 bytes (coll_basic_reduce.c, 194)
Request for 0 bytes (coll_basic_reduce_scatter.c, 80)
Request for 0 bytes (coll_basic_reduce.c, 194)
Request for 0 bytes (coll_basic_reduce_scatter.c, 80)
Request for 0 bytes (coll_basic_reduce.c, 194)

Here is the output of ompi_info command

[root_at_micrompi-1 SRC_PMB]# ompi_info
                Open MPI: 1.0a1r6911
   Open MPI SVN revision: r6911
                Open RTE: 1.0a1r6911
   Open RTE SVN revision: r6911
                    OPAL: 1.0a1r6911
       OPAL SVN revision: r6911
                  Prefix: /openmpi
 Configured architecture: x86_64-redhat-linux-gnu
           Configured by: root
           Configured on: Wed Aug 17 23:59:09 IST 2005
          Configure host: micrompi-1
                Built by: root
                Built on: Thu Aug 18 00:09:33 IST 2005
              Built host: micrompi-1
              C bindings: yes
            C++ bindings: yes
      Fortran77 bindings: yes (all)
      Fortran90 bindings: no
              C compiler: gcc
     C compiler absolute: /usr/bin/gcc
            C++ compiler: g++
   C++ compiler absolute: /usr/bin/g++
      Fortran77 compiler: g77
  Fortran77 compiler abs: /usr/bin/g77
      Fortran90 compiler: none
  Fortran90 compiler abs: none
             C profiling: yes
           C++ profiling: yes
     Fortran77 profiling: yes
     Fortran90 profiling: no
          C++ exceptions: no
          Thread support: posix (mpi: no, progress: no)
  Internal debug support: yes
     MPI parameter check: runtime
Memory profiling support: yes
Memory debugging support: yes
     Memory hook support: yes
         libltdl support: 1
              MCA memory: malloc_hooks (MCA v1.0, API v1.0, Component
v1.0)
           MCA paffinity: linux (MCA v1.0, API v1.0, Component v1.0)
           MCA allocator: basic (MCA v1.0, API v1.0, Component v1.0)
           MCA allocator: bucket (MCA v1.0, API v1.0, Component v1.0)
                MCA coll: basic (MCA v1.0, API v1.0, Component v1.0)
                MCA coll: self (MCA v1.0, API v1.0, Component v1.0)
                  MCA io: romio (MCA v1.0, API v1.0, Component v1.0)
               MCA mpool: mvapi (MCA v1.0, API v1.0, Component v1.0)
               MCA mpool: sm (MCA v1.0, API v1.0, Component v1.0)
                 MCA pml: ob1 (MCA v1.0, API v1.0, Component v1.0)
                 MCA ptl: self (MCA v1.0, API v1.0, Component v1.0)
                 MCA ptl: sm (MCA v1.0, API v1.0, Component v1.0)
                 MCA ptl: tcp (MCA v1.0, API v1.0, Component v1.0)
                 MCA btl: mvapi (MCA v1.0, API v1.0, Component v1.0)
                 MCA btl: self (MCA v1.0, API v1.0, Component v1.0)
                 MCA btl: sm (MCA v1.0, API v1.0, Component v1.0)
                 MCA btl: tcp (MCA v1.0, API v1.0, Component v1.0)
                MCA topo: unity (MCA v1.0, API v1.0, Component v1.0)
                 MCA gpr: null (MCA v1.0, API v1.0, Component v1.0)
                 MCA gpr: proxy (MCA v1.0, API v1.0, Component v1.0)
                 MCA gpr: replica (MCA v1.0, API v1.0, Component v1.0)
                 MCA iof: proxy (MCA v1.0, API v1.0, Component v1.0)
                 MCA iof: svc (MCA v1.0, API v1.0, Component v1.0)
                  MCA ns: proxy (MCA v1.0, API v1.0, Component v1.0)
                  MCA ns: replica (MCA v1.0, API v1.0, Component v1.0)
                 MCA oob: tcp (MCA v1.0, API v1.0, Component v1.0)
                 MCA ras: host (MCA v1.0, API v1.0, Component v1.0)
                 MCA rds: hostfile (MCA v1.0, API v1.0, Component v1.0)
                 MCA rds: resfile (MCA v1.0, API v1.0, Component v1.0)
               MCA rmaps: round_robin (MCA v1.0, API v1.0, Component
v1.0)
                MCA rmgr: proxy (MCA v1.0, API v1.0, Component v1.0)
                MCA rmgr: urm (MCA v1.0, API v1.0, Component v1.0)
                 MCA rml: oob (MCA v1.0, API v1.0, Component v1.0)
                 MCA pls: fork (MCA v1.0, API v1.0, Component v1.0)
                 MCA pls: proxy (MCA v1.0, API v1.0, Component v1.0)
                 MCA pls: rsh (MCA v1.0, API v1.0, Component v1.0)
                 MCA sds: env (MCA v1.0, API v1.0, Component v1.0)
                 MCA sds: pipe (MCA v1.0, API v1.0, Component v1.0)
                 MCA sds: seed (MCA v1.0, API v1.0, Component v1.0)
                 MCA sds: singleton (MCA v1.0, API v1.0, Component v1.0)

Here is the output of sample MPI program which sends a char and recvs a
char.

[root_at_micrompi-1 ~]# mpirun -np 2 ./a.out
Could not join a running, existing universe
Establishing a new one named: default-universe-12913
[0,0,0] mca_oob_tcp_init: calling orte_gpr.subscribe
[0,0,0] mca_oob_tcp_init: calling orte_gpr.put(orte-job-0)
[0,0,0] mca_oob_tcp_accept: 10.52.0.98:52866
[0,0,0]-[0,0,1] accepted: 10.52.0.98 - 10.52.0.98 nodelay 0 sndbuf 50724
rcvbuf 87552 flags 00000802
[0,0,0]-[0,0,1] mca_oob_tcp_send: tag 2
[0,0,0]-[0,0,1] mca_oob_tcp_send: tag 2
[0,0,0]-[0,0,1] mca_oob_tcp_send: tag 2
[0,0,0]-[0,0,1] mca_oob_tcp_send: tag 2
[0,0,0]-[0,0,1] mca_oob_tcp_send: tag 2
[0,0,0]-[0,0,1] mca_oob_tcp_send: tag 2
[0,1,0] mca_oob_tcp_init: calling orte_gpr.subscribe
[0,1,0] mca_oob_tcp_init: calling orte_gpr.put(orte-job-1)
[0,0,0] mca_oob_tcp_accept: 10.52.0.98:53634
[0,1,0]-[0,0,0] mca_oob_tcp_send: tag 2
[0,1,0]-[0,0,0] mca_oob_tcp_peer_start_connect: connecting port 33487
to: 10.52.0.98:33483
[0,0,0]-[0,1,0] accepted: 10.52.0.98 - 10.52.0.98 nodelay 0 sndbuf 50724
rcvbuf 87552 flags 00000802
[0,1,0]-[0,0,0] connected: 10.52.0.98 - 10.52.0.98 nodelay 0 sndbuf
50724 rcvbuf 87552 flags 00000802
[0,0,0]-[0,1,0] mca_oob_tcp_send: tag 2
[0,1,0]-[0,0,0] mca_oob_tcp_recv: tag 2
[0,1,0]-[18446744073709551615,18446744073709551615,18446744073709551615]
mca_oob_tcp_recv: tag 10
[0,0,0] mca_oob_tcp_accept: 10.52.0.97:13953
[0,0,0]-[0,0,2] accepted: 10.52.0.98 - 10.52.0.97 nodelay 0 sndbuf 16384
rcvbuf 87380 flags 00000802
[0,0,0]-[0,0,2] mca_oob_tcp_send: tag 2
[0,0,0]-[0,0,2] mca_oob_tcp_send: tag 2
[0,0,0]-[0,0,2] mca_oob_tcp_send: tag 2
[0,0,0]-[0,0,2] mca_oob_tcp_send: tag 2
[0,0,0]-[0,0,2] mca_oob_tcp_send: tag 2
[0,0,0]-[0,0,2] mca_oob_tcp_send: tag 2
[0,1,1] mca_oob_tcp_init: calling orte_gpr.subscribe
[0,1,1] mca_oob_tcp_init: calling orte_gpr.put(orte-job-1)
[0,1,1]-[0,0,0] mca_oob_tcp_send: tag 2
[0,0,0] mca_oob_tcp_accept: 10.52.0.97:14721
[0,1,1]-[0,0,0] mca_oob_tcp_peer_start_connect: connecting port 33079
to: 10.52.0.98:33483
[0,0,0]-[0,1,1] accepted: 10.52.0.98 - 10.52.0.97 nodelay 0 sndbuf 16384
rcvbuf 87380 flags 00000802
[0,1,1]-[0,0,0] connected: 10.52.0.97 - 10.52.0.98 nodelay 0 sndbuf
16384 rcvbuf 87380 flags 00000802
[0,0,0]-[0,1,1] mca_oob_tcp_send: tag 2
[0,0,0]-[0,1,1] mca_oob_tcp_send: tag 3
[0,0,0]-[0,1,0] mca_oob_tcp_send: tag 3
[0,0,0]-[0,1,0] mca_oob_tcp_send: tag 10
[0,0,0]-[0,1,1] mca_oob_tcp_send: tag 10
[0,1,0] mca_oob_tcp_registry_callback
[0,1,1]-[0,0,0] mca_oob_tcp_recv: tag 2
[0,1,0] mca_oob_tcp_registry_callback: received peer [0,1,0]
[0,1,0] mca_oob_tcp_registry_callback: received peer [0,1,1]
[0,1,1]-[18446744073709551615,18446744073709551615,18446744073709551615]
mca_oob_tcp_recv: tag 10
[0,1,1] mca_oob_tcp_registry_callback
[0,1,1] mca_oob_tcp_registry_callback: received peer [0,1,0]
[0,1,1] mca_oob_tcp_registry_callback: received peer [0,1,1]
[0,0,0]-[0,1,1] mca_oob_tcp_send: tag 2
[0,1,1]-[0,0,0] mca_oob_tcp_send: tag 2
[0,1,1]-[0,0,0] mca_oob_tcp_recv: tag 2
[0,0,0]-[0,1,1] mca_oob_tcp_send: tag 2
[0,1,1]-[0,0,0] mca_oob_tcp_send: tag 2
[0,1,1]-[0,0,0] mca_oob_tcp_recv: tag 2
[0,0,0]-[0,1,0] mca_oob_tcp_send: tag 2
[0,1,0]-[0,0,0] mca_oob_tcp_send: tag 2
[0,1,0]-[0,0,0] mca_oob_tcp_recv: tag 2
[0,0,0]-[0,1,0] mca_oob_tcp_send: tag 2
[0,0,0]-[0,1,0] mca_oob_tcp_send: tag 10
[0,0,0]-[0,1,1] mca_oob_tcp_send: tag 10
[0,1,1]-[18446744073709551615,18446744073709551615,18446744073709551615]
mca_oob_tcp_recv: tag 10
[0,1,0]-[0,0,0] mca_oob_tcp_send: tag 2
[0,1,0]-[0,0,0] mca_oob_tcp_recv: tag 2
[0,1,0]-[18446744073709551615,18446744073709551615,18446744073709551615]
mca_oob_tcp_recv: tag 10
[0,1,0]-[0,1,1] mca_oob_tcp_peer_start_connect: connecting port 33487
to: 10.52.0.97:33079
[0,1,1] mca_oob_tcp_accept: 10.52.0.98:53890
[0,1,0]-[0,1,1] connected: 10.52.0.98 - 10.52.0.97 nodelay 0 sndbuf
16384 rcvbuf 87380 flags 00000802
[0,1,1]-[0,1,0] accepted: 10.52.0.97 - 10.52.0.98 nodelay 0 sndbuf 16384
rcvbuf 87380 flags 00000802
Task 1: Received 1 char(s) from task 0 with tag 1
[0,0,0]-[0,1,1] mca_oob_tcp_send: tag 2
Task 0: Received 1 char(s) from task 1 with tag 1
[0,0,0]-[0,1,0] mca_oob_tcp_send: tag 2
[0,1,0]-[0,0,0] mca_oob_tcp_send: tag 2
[0,1,1]-[0,0,0] mca_oob_tcp_send: tag 2
[0,1,0]-[0,0,0] mca_oob_tcp_recv: tag 2
[0,0,0]-[0,1,0] mca_oob_tcp_send: tag 2
[0,1,1]-[0,0,0] mca_oob_tcp_recv: tag 2
[0,0,0]-[0,1,1] mca_oob_tcp_send: tag 2
[0,0,0]-[0,1,0] mca_oob_tcp_send: tag 10
[0,0,0]-[0,1,1] mca_oob_tcp_send: tag 10
[0,1,0]-[0,0,0] mca_oob_tcp_send: tag 2
[0,1,1]-[0,0,0] mca_oob_tcp_send: tag 2
[0,1,0]-[0,0,0] mca_oob_tcp_recv: tag 2
[0,1,1]-[0,0,0] mca_oob_tcp_recv: tag 2
[0,1,0]-[18446744073709551615,18446744073709551615,18446744073709551615]
mca_oob_tcp_recv: tag 10
[0,1,1]-[18446744073709551615,18446744073709551615,18446744073709551615]
mca_oob_tcp_recv: tag 10
[0,1,0]-[0,0,0] mca_oob_tcp_send: tag 2
[0,0,0]-[0,1,0] mca_oob_tcp_send: tag 2
[0,1,0]-[0,0,0] mca_oob_tcp_recv: tag 2
[0,1,1]-[0,0,0] mca_oob_tcp_send: tag 2
[0,0,0]-[0,1,1] mca_oob_tcp_send: tag 2
[0,1,1]-[0,0,0] mca_oob_tcp_recv: tag 2
[0,1,0]-[0,0,0] mca_oob_tcp_send: tag 2
[0,0,0]-[0,1,0] mca_oob_tcp_send: tag 2
[0,1,0]-[0,0,0] mca_oob_tcp_recv: tag 2
[0,0,0]-[0,1,1] mca_oob_tcp_send: tag 2
[0,0,0]-[0,1,0] mca_oob_tcp_send: tag 10
[0,0,0]-[0,1,1] mca_oob_tcp_send: tag 10
[0,1,0]-[18446744073709551615,18446744073709551615,18446744073709551615]
mca_oob_tcp_recv: tag 10
[0,1,1]-[0,0,0] mca_oob_tcp_send: tag 2
[0,1,1]-[0,0,0] mca_oob_tcp_recv: tag 2
[0,1,1]-[18446744073709551615,18446744073709551615,18446744073709551615]
mca_oob_tcp_recv: tag 10
[0,0,0]-[0,1,0] mca_oob_tcp_msg_recv: peer closed connection
[0,0,0]-[0,1,0] mca_oob_tcp_peer_close(0x526810) sd 6 state 4
[0,1,1]-[0,1,0] mca_oob_tcp_msg_recv: peer closed connection
[0,1,1]-[0,1,0] mca_oob_tcp_peer_close(0x521210) sd 13 state 4
[0,0,0]-[0,1,1] mca_oob_tcp_msg_recv: peer closed connection
[0,0,0]-[0,1,1] mca_oob_tcp_peer_close(0x526e30) sd 8 state 4
[0,0,0]-[0,0,2] mca_oob_tcp_send: tag 2
[0,0,0]-[0,0,2] mca_oob_tcp_send: tag 2
[0,0,0]-[0,0,2] mca_oob_tcp_msg_recv: peer closed connection
[0,0,0]-[0,0,2] mca_oob_tcp_peer_close(0x526b20) sd 7 state 4
[0,0,0]-[0,0,1] mca_oob_tcp_send: tag 2
[0,0,0]-[0,0,1] mca_oob_tcp_send: tag 2

My configure command looks like

./configure --prefix=/openmpi --with-btl-mvapi=/usr/local/topspin/
--enable-mca-no-build=btl-openib,pml-teg,pml-uniq

Since I am working with mvapi component, I disabled openib. But I could
see that data is going over TCP/GigE and not on Infiniband.

I have run pallas, it simply hangs again :-(

Note: I added pml=ob1 in the conf file
/openmpi/etc/openmpi-mca-params.conf

Any latest options being added to the configure command? Please let me
know.

Thanks
-Sridhar

-----Original Message-----
From: devel-bounces_at_[hidden] [mailto:devel-bounces_at_[hidden]] On
Behalf Of George Bosilca
Sent: Wednesday, August 10, 2005 11:56 PM
To: Open MPI Developers
Subject: Re: [O-MPI devel] Fwd: Regarding MVAPI Component in Open MPI

I just commit the collective fix. Now I'm able to run the ex-pallas
test on 2, 4 and 8 nodes without any errors. However, I'm unable to
test the IB BTL as I don't have access to any machine having such
hardware.

   george.

PS: if one still notice the malloc with 0 bytes length, please sent a
bug report.

On Aug 10, 2005, at 6:28 AM, Sridhar Chirravuri wrote:

>
> Hi,
>
> I got the latest code drop of 6791 today morning.
>
> I have removed .ompi_ignore and .ompi_unignore files from
> ompi/mca/mpool/mvapi directory. If I don't remove and build, the MPI
> program fails with signal 11. After removing those hidden files from
> that directory and building, signal 11 error disappeared.
>
> I have configured with the options given by Galen.
>
> ./configure --prefix=/openmpi --with-btl-mvapi=/usr/local/topspin/
> --enable-mca-no-build=btl-openib,pml-teg,pml-uniq
>
> After make all install, I have run pallas but I got the same error
> messages (please see down below for error messages). I have run 3-4
> times, sometimes I didn't get any output but pallas just hungs. I have
> run pingpong only. I have run pallas (all functions including reduce),
> but got the following messages in intra-node case.
>
> Request for 0 bytes (coll_basic_reduce_scatter.c, 79)
> Request for 0 bytes (coll_basic_reduce.c, 193)
> Request for 0 bytes (coll_basic_reduce_scatter.c, 79)
> Request for 0 bytes (coll_basic_reduce.c, 193)
>
> Since these types of messages seen by George, upcoming patch might
> resolve this issue.
>
> Also, I have run mpi-ping.c program given by Galen with the latest
> code
> drop and it just hung. Here is the output
>
> [root_at_micrompi-1 ~]# mpirun -np 2 ./a.out -r 10 0 100000 1000
> Could not join a running, existing universe
> Establishing a new one named: default-universe-12461
> mpi-ping: ping-pong
> nprocs=2, reps=10, min bytes=0, max bytes=100000 inc bytes=1000
> 0 pings 1
>
>
> ... I just did ctrl+c here after 10 mins ...
>
> 2 processes killed (possibly by Open MPI)
>
> I have no clue whether the George patch will fix this problem or not.
>
> Before running mpi-ping program, I have export
> OMPI_MCA_btl_base_debug=2
> in my shell.
>
> Thanks
> -Sridhar
>
> -----Original Message-----
> From: devel-bounces_at_[hidden] [mailto:devel-bounces_at_open-
> mpi.org] On
> Behalf Of Galen Shipman
> Sent: Tuesday, August 09, 2005 11:10 PM
> To: Open MPI Developers
> Subject: Re: [O-MPI devel] Fwd: Regarding MVAPI Component in Open MPI
>
> Hi
> On Aug 9, 2005, at 8:15 AM, Sridhar Chirravuri wrote:
>
>
>
>> The same kind of output while running Pallas "pingpong" test.
>>
>> -Sridhar
>>
>> -----Original Message-----
>> From: devel-bounces_at_[hidden] [mailto:devel-bounces_at_[hidden]]
>>
>>
> On
>
>
>> Behalf Of Sridhar Chirravuri
>> Sent: Tuesday, August 09, 2005 7:44 PM
>> To: Open MPI Developers
>> Subject: Re: [O-MPI devel] Fwd: Regarding MVAPI Component in Open MPI
>>
>>
>> I have run sendrecv function in Pallas but it failed to run it. Here
>>
>>
> is
>
>
>> the output
>>
>> [root_at_micrompi-2 SRC_PMB]# mpirun -np 2 PMB-MPI1 sendrecv
>> Could not join a running, existing universe
>> Establishing a new one named: default-universe-5097
>> [0,1,1][btl_mvapi.c:130:mca_btl_mvapi_del_procs] Stub
>> [0,1,1][btl_mvapi.c:130:mca_btl_mvapi_del_procs] Stub
>>
>>
>> [0,1,0][btl_mvapi.c:130:mca_btl_mvapi_del_procs] Stub
>>
>> [0,1,0][btl_mvapi.c:130:mca_btl_mvapi_del_procs] Stub
>>
>> [0,1,0][btl_mvapi_endpoint.c:542:mca_btl_mvapi_endpoint_send]
>> Connection
>> to endpoint closed ... connecting ...
>> [0,1,0][btl_mvapi_endpoint.c:
>> 318:mca_btl_mvapi_endpoint_start_connect]
>> Initialized High Priority QP num = 263177, Low Priority QP num =
>> 263178,
>> LID = 785
>>
>> [0,1,0][btl_mvapi_endpoint.c:190:
>> mca_btl_mvapi_endpoint_send_connect_req
>> ] Sending High Priority QP num = 263177, Low Priority QP num =
>> 263178,
>> LID = 785[0,1,0][btl_mvapi_endpoint.c:
>> 542:mca_btl_mvapi_endpoint_send]
>> Connection to endpoint closed ... connecting ...
>> [0,1,0][btl_mvapi_endpoint.c:
>> 318:mca_btl_mvapi_endpoint_start_connect]
>> Initialized High Priority QP num = 263179, Low Priority QP num =
>> 263180,
>> LID = 786
>>
>> [0,1,0][btl_mvapi_endpoint.c:190:
>> mca_btl_mvapi_endpoint_send_connect_req
>> ] Sending High Priority QP num = 263179, Low Priority QP num =
>> 263180,
>> LID = 786#---------------------------------------------------
>> # PALLAS MPI Benchmark Suite V2.2, MPI-1 part
>> #---------------------------------------------------
>> # Date : Tue Aug 9 07:11:25 2005
>> # Machine : x86_64# System : Linux
>> # Release : 2.6.9-5.ELsmp
>> # Version : #1 SMP Wed Jan 5 19:29:47 EST 2005
>>
>> #
>> # Minimum message length in bytes: 0
>> # Maximum message length in bytes: 4194304
>> #
>> # MPI_Datatype : MPI_BYTE
>> # MPI_Datatype for reductions : MPI_FLOAT
>> # MPI_Op : MPI_SUM
>> #
>> #
>>
>> # List of Benchmarks to run:
>>
>> # Sendrecv
>> [0,1,1][btl_mvapi_endpoint.c:368:
>> mca_btl_mvapi_endpoint_reply_start_conn
>> ect] Initialized High Priority QP num = 263177, Low Priority QP num =
>> 263178, LID = 777
>>
>> [0,1,1][btl_mvapi_endpoint.c:266:
>> mca_btl_mvapi_endpoint_set_remote_info]
>> Received High Priority QP num = 263177, Low Priority QP num 263178,
>> LID
>> = 785
>>
>> [0,1,1][btl_mvapi_endpoint.c:
>> 756:mca_btl_mvapi_endpoint_qp_init_query]
>> Modified to init..Qp
>> 7080096[0,1,1][btl_mvapi_endpoint.c:791:
>> mca_btl_mvapi_endpoint_qp_init_q
>> uery] Modified to RTR..Qp
>> 7080096[0,1,1][btl_mvapi_endpoint.c:814:
>> mca_btl_mvapi_endpoint_qp_init_q
>> uery] Modified to RTS..Qp 7080096
>>
>> [0,1,1][btl_mvapi_endpoint.c:
>> 756:mca_btl_mvapi_endpoint_qp_init_query]
>> Modified to init..Qp 7240736
>> [0,1,1][btl_mvapi_endpoint.c:
>> 791:mca_btl_mvapi_endpoint_qp_init_query]
>> Modified to RTR..Qp
>> 7240736[0,1,1][btl_mvapi_endpoint.c:814:
>> mca_btl_mvapi_endpoint_qp_init_q
>> uery] Modified to RTS..Qp 7240736
>> [0,1,1][btl_mvapi_endpoint.c:190:
>> mca_btl_mvapi_endpoint_send_connect_req
>> ] Sending High Priority QP num = 263177, Low Priority QP num =
>> 263178,
>> LID = 777
>> [0,1,0][btl_mvapi_endpoint.c:266:
>> mca_btl_mvapi_endpoint_set_remote_info]
>> Received High Priority QP num = 263177, Low Priority QP num 263178,
>> LID
>> = 777
>> [0,1,0][btl_mvapi_endpoint.c:
>> 756:mca_btl_mvapi_endpoint_qp_init_query]
>> Modified to init..Qp 7081440
>> [0,1,0][btl_mvapi_endpoint.c:
>> 791:mca_btl_mvapi_endpoint_qp_init_query]
>> Modified to RTR..Qp 7081440
>> [0,1,0][btl_mvapi_endpoint.c:
>> 814:mca_btl_mvapi_endpoint_qp_init_query]
>> Modified to RTS..Qp 7081440
>> [0,1,0][btl_mvapi_endpoint.c:
>> 756:mca_btl_mvapi_endpoint_qp_init_query]
>> Modified to init..Qp 7241888
>> [0,1,0][btl_mvapi_endpoint.c:
>> 791:mca_btl_mvapi_endpoint_qp_init_query]
>> Modified to RTR..Qp
>> 7241888[0,1,0][btl_mvapi_endpoint.c:814:
>> mca_btl_mvapi_endpoint_qp_init_q
>> uery] Modified to RTS..Qp 7241888
>> [0,1,1][btl_mvapi_component.c:523:mca_btl_mvapi_component_progress]
>>
>>
> Got
>
>
>> a recv completion
>>
>>
>> Thanks
>> -Sridhar
>>
>>
>>
>>
>> -----Original Message-----
>> From: devel-bounces_at_[hidden] [mailto:devel-bounces_at_[hidden]]
>>
>>
> On
>
>
>> Behalf Of Brian Barrett
>> Sent: Tuesday, August 09, 2005 7:35 PM
>> To: Open MPI Developers
>> Subject: Re: [O-MPI devel] Fwd: Regarding MVAPI Component in Open MPI
>>
>> On Aug 9, 2005, at 8:48 AM, Sridhar Chirravuri wrote:
>>
>>
>>
>>> Does r6774 has lot of changes that are related to 3rd generation
>>> point-to-point? I am trying to run some benchmark tests (ex:
>>> pallas) with Open MPI stack and just want to compare the
>>> performance figures with MVAPICH 095 and MVAPICH 092.
>>>
>>> In order to use 3rd generation p2p communication, I have added the
>>> following line in the /openmpi/etc/openmpi-mca-params.conf
>>>
>>> pml=ob1
>>>
>>> I also exported (as double check) OMPI_MCA_pml=ob1.
>>>
>>> Then, I have tried running on the same machine. My machine has got
>>> 2 processors.
>>>
>>> Mpirun -np 2 ./PMB-MPI1
>>>
>>> I still see the following lines
>>>
>>> Request for 0 bytes (coll_basic_reduce_scatter.c, 79)
>>> Request for 0 bytes (coll_basic_reduce.c, 193)
>>> Request for 0 bytes (coll_basic_reduce_scatter.c, 79)
>>> Request for 0 bytes (coll_basic_reduce.c, 193)
>>>
>>>
>>
>> These errors are coming from the collective routines, not the PML/BTL
>> layers. It looks like the reduction codes are trying to call malloc
>> (0), which doesn't work so well. We'll take a look as soon as we
>> can. In the mean time, can you just not run the tests that call the
>> reduction collectives?
>>
>> Brian
>>
>>
>> --
>> Brian Barrett
>> Open MPI developer
>> http://www.open-mpi.org/
>>
>>
>> _______________________________________________
>> devel mailing list
>> devel_at_[hidden]
>> http://www.open-mpi.org/mailman/listinfo.cgi/devel
>>
>> _______________________________________________
>> devel mailing list
>> devel_at_[hidden]
>> http://www.open-mpi.org/mailman/listinfo.cgi/devel
>>
>> _______________________________________________
>> devel mailing list
>> devel_at_[hidden]
>> http://www.open-mpi.org/mailman/listinfo.cgi/devel
>>
>>
>
> _______________________________________________
> devel mailing list
> devel_at_[hidden]
> http://www.open-mpi.org/mailman/listinfo.cgi/devel
>
> _______________________________________________
> devel mailing list
> devel_at_[hidden]
> http://www.open-mpi.org/mailman/listinfo.cgi/devel
>
>

_______________________________________________
devel mailing list
devel_at_[hidden]
http://www.open-mpi.org/mailman/listinfo.cgi/devel

_______________________________________________
devel mailing list
devel_at_[hidden]
http://www.open-mpi.org/mailman/listinfo.cgi/devel