I got a problem with my code, wich run some kinf of a simulator.
I get 4 worker (aka 4 mpi process ) wich process data.
These data aren't available at the same time, so i get another process
(Splitter) wich send chunk of data to each process in round robin.
This work well using MPI_Send and Receive but aftet that i need to
reduce all the data.
I hope to be able to use MPI_Reduce to reduce all data from all worker
but there is a problem :
1. All results data aren't available at the same time, dut to the delay
from the original data delay.
2. I cannot wait all data to be computed, i need to proceed the reduce a
soon as possible
So when the first and second worker have finished, i can reduce the two
results data and keep the results on the first worker.
And when the third and the fourth have finished, i can reduce these two
too, and keep results on third worker.
At last i need to reduce data from first and third worker.
The only way to do that using MPI_Reduce is to create "communicators".
All i want is :
commA : contain W1 W2
commB : contain W3 W4
commC : contain W1 W3
Let's say i've already create a communicator only for my workers
I can easily add this line in all my workers :
*If i understand well i will get an communicator on W1 and W2 wich
contains W1 and W2, and a communicator on W3 and W4 wich contains W3 and
W4. Am i right?*
But next when i try to use (only on W1 and W3):
*I need also to call MPI_Create_comm on W2 and W4 or it will block. Why?*
After that, i got lot of non persistent error depending of the number of
worker i want to use.
*So is this allowed to create and use overlapping communicator? and if
so how to do that?*