Open MPI logo

Open MPI Development Mailing List Archives

  |   Home   |   Support   |   FAQ   |   all Development mailing list

Subject: Re: [OMPI devel] Pallas fails
From: Rolf vandeVaart (Rolf.Vandevaart_at_[hidden])
Date: 2008-06-12 09:16:37


We are seeing the same assert errors. Not sure when they started. If
you run the tests without the sm btl, then they pass..

I have opened a ticket at https://svn.open-mpi.org/trac/ompi/ticket/1344

If anyone has any other info, add it to the ticket.

Rolf

Pavel Shamis (Pasha) wrote:
> With 1.3a1r18643 Pallas tests pass on my machine.
> But I see new failures (assertion) in Intel-Test
> http://www.open-mpi.org/mtt/index.php?do_redir=733
>
> PI_Type_struct_types_c: btl_sm.c:684: mca_btl_sm_sendi: Assertion `max_data == payload_size'
> failed.
> [sw216:32013] *** Process received signal ***
> [sw216:32013] Signal: Aborted (6)
> [sw216:32013] Signal code: (-6)
> [sw216:32013] [ 0] /lib64/libpthread.so.0 [0x2aba5e51ec10]
> [sw216:32013] [ 1] /lib64/libc.so.6(gsignal+0x35) [0x2aba5e657b95]
> [sw216:32013] [ 2] /lib64/libc.so.6(abort+0x110) [0x2aba5e658f90]
> [sw216:32013] [ 3] /lib64/libc.so.6(__assert_fail+0xf6) [0x2aba5e651256]
> [sw216:32013] [ 4]
>
>
>
>
> Pavel Shamis (Pasha) wrote:
>
>> Last conf. call Jeff mentioned that he see some collectives failures.
>> In my MTT testing I also see that Pallas collectives failed - http://www.open-mpi.org/mtt/index.php?do_redir=682
>>
>>
>> Alltoall
>>
>> #----------------------------------------------------------------
>> # Benchmarking Alltoall
>> # #processes = 20
>> #----------------------------------------------------------------
>> #bytes #repetitions t_min[usec] t_max[usec] t_avg[usec]
>> 0 1000 0.03 0.05 0.04
>> 1 1000 179.15 179.22 179.18
>> 2 1000 155.96 156.02 155.98
>> 4 1000 156.93 156.98 156.95
>> 8 1000 163.63 163.67 163.65
>> 16 1000 115.04 115.08 115.07
>> 32 1000 123.57 123.62 123.59
>> 64 1000 129.78 129.82 129.80
>> 128 1000 141.45 141.49 141.48
>> 256 1000 960.11 960.24 960.20
>> 512 1000 900.95 901.11 901.04
>> 1024 1000 921.95 922.05 922.00
>> 2048 1000 862.50 862.72 862.60
>> 4096 1000 1044.90 1044.95 1044.92
>> 8192 1000 1458.59 1458.77 1458.69
>> *** An error occurred in MPI_Alltoall
>> *** on communicator MPI COMMUNICATOR 4 SPLIT FROM 0
>> *** An error occurred in MPI_Alltoall
>> *** on communicator MPI COMMUNICATOR 4 SPLIT FROM 0
>>
>> _______________________________________________
>> devel mailing list
>> devel_at_[hidden]
>> http://www.open-mpi.org/mailman/listinfo.cgi/devel
>>
>>
>>
>
> _______________________________________________
> devel mailing list
> devel_at_[hidden]
> http://www.open-mpi.org/mailman/listinfo.cgi/devel
>