Open MPI logo

Open MPI User's Mailing List Archives

  |   Home   |   Support   |   FAQ   |   all Open MPI User's mailing list

Subject: Re: [OMPI users] MPI_ERR_IN_STATUS from MPI_Bcast?
From: Jeff Squyres (jsquyres_at_[hidden])
Date: 2011-02-14 13:11:06


Thanks Jeremiah; I filed the following ticket about this:

    https://svn.open-mpi.org/trac/ompi/ticket/2723

On Feb 10, 2011, at 3:24 PM, Jeremiah Willcock wrote:

> I forgot to mention that this was tested with 3 or 4 ranks, connected via TCP.
>
> -- Jeremiah Willcock
>
> On Thu, 10 Feb 2011, Jeremiah Willcock wrote:
>
>> Here is a small test case that hits the bug on 1.4.1:
>>
>> #include <mpi.h>
>>
>> int arr[1142];
>>
>> int main(int argc, char** argv) {
>> int rank, my_size;
>> MPI_Init(&argc, &argv);
>> MPI_Comm_rank(MPI_COMM_WORLD, &rank);
>> my_size = (rank == 1) ? 1142 : 1088;
>> MPI_Bcast(arr, my_size, MPI_INT, 0, MPI_COMM_WORLD);
>> MPI_Finalize();
>> return 0;
>> }
>>
>> I tried it on 1.5.1, and I get MPI_ERR_TRUNCATE instead, so this might have already been fixed.
>>
>> -- Jeremiah Willcock
>>
>>
>> On Thu, 10 Feb 2011, Jeremiah Willcock wrote:
>>
>>> FYI, I am having trouble finding a small test case that will trigger this on 1.5; I'm either getting deadlocks or MPI_ERR_TRUNCATE, so it could have been fixed. What are the triggering rules for different broadcast algorithms? It could be that only certain sizes or only certain BTLs trigger it.
>>> -- Jeremiah Willcock
>>> On Thu, 10 Feb 2011, Jeff Squyres wrote:
>>>> Nifty! Yes, I agree that that's a poor error message. It's probably (unfortunately) being propagated up from the underlying point-to-point system, where an ERR_IN_STATUS would actually make sense.
>>>> I'll file a ticket about this. Thanks for the heads up.
>>>> On Feb 9, 2011, at 4:49 PM, Jeremiah Willcock wrote:
>>>>> On Wed, 9 Feb 2011, Jeremiah Willcock wrote:
>>>>>> I get the following Open MPI error from 1.4.1:
>>>>>> *** An error occurred in MPI_Bcast
>>>>>> *** on communicator MPI COMMUNICATOR 3 SPLIT FROM 0
>>>>>> *** MPI_ERR_IN_STATUS: error code in status
>>>>>> *** MPI_ERRORS_ARE_FATAL (your MPI job will now abort)
>>>>>> (hostname and port removed from each line). There is no MPI_Status returned by MPI_Bcast, so I don't know what the error is? Is this something that people have seen before?
>>>>> For the record, this appears to be caused by specifying inconsistent data sizes on the different ranks in the broadcast operation. The error message could still be improved, though.
>>>>> -- Jeremiah Willcock
>>>>> _______________________________________________
>>>>> users mailing list
>>>>> users_at_[hidden]
>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>> --
>>>> Jeff Squyres
>>>> jsquyres_at_[hidden]
>>>> For corporate legal information go to:
>>>> http://www.cisco.com/web/about/doing_business/legal/cri/
>>>> _______________________________________________
>>>> users mailing list
>>>> users_at_[hidden]
>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>> _______________________________________________
>>> users mailing list
>>> users_at_[hidden]
>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>> _______________________________________________
>> users mailing list
>> users_at_[hidden]
>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>
> _______________________________________________
> users mailing list
> users_at_[hidden]
> http://www.open-mpi.org/mailman/listinfo.cgi/users

-- 
Jeff Squyres
jsquyres_at_[hidden]
For corporate legal information go to:
http://www.cisco.com/web/about/doing_business/legal/cri/